var/home/core/zuul-output/0000755000175000017500000000000015071402134014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071406742015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004454517615071406733017723 0ustar rootrootOct 08 06:31:36 crc systemd[1]: Starting Kubernetes Kubelet... Oct 08 06:31:36 crc restorecon[4737]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:36 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 06:31:37 crc restorecon[4737]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 08 06:31:37 crc kubenswrapper[4810]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 06:31:37 crc kubenswrapper[4810]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 08 06:31:37 crc kubenswrapper[4810]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 06:31:37 crc kubenswrapper[4810]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 06:31:37 crc kubenswrapper[4810]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 08 06:31:37 crc kubenswrapper[4810]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.812157 4810 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815603 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815620 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815624 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815629 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815633 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815637 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815642 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815648 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815652 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815657 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815662 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815667 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815671 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815676 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815679 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815684 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815695 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815700 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815705 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815710 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815714 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815718 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815722 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815725 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815729 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815732 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815736 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815739 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815743 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815748 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815753 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815756 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815760 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815763 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815767 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815771 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815774 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815778 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815782 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815786 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815789 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815794 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815798 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815802 4810 feature_gate.go:330] unrecognized feature gate: Example Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815806 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815810 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815813 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815817 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815820 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815824 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815827 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815830 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815834 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815837 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815840 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815844 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815848 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815851 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815854 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815858 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815861 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815865 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815869 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815872 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815876 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815879 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815883 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815886 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815890 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815894 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.815898 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.815990 4810 flags.go:64] FLAG: --address="0.0.0.0" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.815998 4810 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816006 4810 flags.go:64] FLAG: --anonymous-auth="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816012 4810 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816018 4810 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816023 4810 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816029 4810 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816035 4810 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816039 4810 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816044 4810 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816049 4810 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816053 4810 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816058 4810 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816062 4810 flags.go:64] FLAG: --cgroup-root="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816065 4810 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816070 4810 flags.go:64] FLAG: --client-ca-file="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816073 4810 flags.go:64] FLAG: --cloud-config="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816077 4810 flags.go:64] FLAG: --cloud-provider="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816081 4810 flags.go:64] FLAG: --cluster-dns="[]" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816087 4810 flags.go:64] FLAG: --cluster-domain="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816091 4810 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816095 4810 flags.go:64] FLAG: --config-dir="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816100 4810 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816106 4810 flags.go:64] FLAG: --container-log-max-files="5" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816114 4810 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816120 4810 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816126 4810 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816131 4810 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816136 4810 flags.go:64] FLAG: --contention-profiling="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816141 4810 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816145 4810 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816149 4810 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816154 4810 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816160 4810 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816165 4810 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816170 4810 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816174 4810 flags.go:64] FLAG: --enable-load-reader="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816179 4810 flags.go:64] FLAG: --enable-server="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816183 4810 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816196 4810 flags.go:64] FLAG: --event-burst="100" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816200 4810 flags.go:64] FLAG: --event-qps="50" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816204 4810 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816209 4810 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816213 4810 flags.go:64] FLAG: --eviction-hard="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816219 4810 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816222 4810 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816227 4810 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816231 4810 flags.go:64] FLAG: --eviction-soft="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816235 4810 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816239 4810 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816243 4810 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816247 4810 flags.go:64] FLAG: --experimental-mounter-path="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816251 4810 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816255 4810 flags.go:64] FLAG: --fail-swap-on="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816259 4810 flags.go:64] FLAG: --feature-gates="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816264 4810 flags.go:64] FLAG: --file-check-frequency="20s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816268 4810 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816272 4810 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816277 4810 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816281 4810 flags.go:64] FLAG: --healthz-port="10248" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816285 4810 flags.go:64] FLAG: --help="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816289 4810 flags.go:64] FLAG: --hostname-override="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816293 4810 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816297 4810 flags.go:64] FLAG: --http-check-frequency="20s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816301 4810 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816305 4810 flags.go:64] FLAG: --image-credential-provider-config="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816309 4810 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816314 4810 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816318 4810 flags.go:64] FLAG: --image-service-endpoint="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816322 4810 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816325 4810 flags.go:64] FLAG: --kube-api-burst="100" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816329 4810 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816334 4810 flags.go:64] FLAG: --kube-api-qps="50" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816338 4810 flags.go:64] FLAG: --kube-reserved="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816341 4810 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816346 4810 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816350 4810 flags.go:64] FLAG: --kubelet-cgroups="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816354 4810 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816358 4810 flags.go:64] FLAG: --lock-file="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816362 4810 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816366 4810 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816370 4810 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816376 4810 flags.go:64] FLAG: --log-json-split-stream="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816380 4810 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816384 4810 flags.go:64] FLAG: --log-text-split-stream="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816388 4810 flags.go:64] FLAG: --logging-format="text" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816392 4810 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816397 4810 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816402 4810 flags.go:64] FLAG: --manifest-url="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816406 4810 flags.go:64] FLAG: --manifest-url-header="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816412 4810 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816417 4810 flags.go:64] FLAG: --max-open-files="1000000" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816424 4810 flags.go:64] FLAG: --max-pods="110" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816430 4810 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816434 4810 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816439 4810 flags.go:64] FLAG: --memory-manager-policy="None" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816443 4810 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816447 4810 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816455 4810 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816459 4810 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816469 4810 flags.go:64] FLAG: --node-status-max-images="50" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816472 4810 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816477 4810 flags.go:64] FLAG: --oom-score-adj="-999" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816481 4810 flags.go:64] FLAG: --pod-cidr="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816485 4810 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816492 4810 flags.go:64] FLAG: --pod-manifest-path="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816496 4810 flags.go:64] FLAG: --pod-max-pids="-1" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816500 4810 flags.go:64] FLAG: --pods-per-core="0" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816504 4810 flags.go:64] FLAG: --port="10250" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816508 4810 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816512 4810 flags.go:64] FLAG: --provider-id="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816517 4810 flags.go:64] FLAG: --qos-reserved="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816521 4810 flags.go:64] FLAG: --read-only-port="10255" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816526 4810 flags.go:64] FLAG: --register-node="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816530 4810 flags.go:64] FLAG: --register-schedulable="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816534 4810 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816540 4810 flags.go:64] FLAG: --registry-burst="10" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816544 4810 flags.go:64] FLAG: --registry-qps="5" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816548 4810 flags.go:64] FLAG: --reserved-cpus="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816552 4810 flags.go:64] FLAG: --reserved-memory="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816557 4810 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816561 4810 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816565 4810 flags.go:64] FLAG: --rotate-certificates="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816569 4810 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816573 4810 flags.go:64] FLAG: --runonce="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816578 4810 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816582 4810 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816588 4810 flags.go:64] FLAG: --seccomp-default="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816592 4810 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816596 4810 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816600 4810 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816606 4810 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816610 4810 flags.go:64] FLAG: --storage-driver-password="root" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816614 4810 flags.go:64] FLAG: --storage-driver-secure="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816619 4810 flags.go:64] FLAG: --storage-driver-table="stats" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816623 4810 flags.go:64] FLAG: --storage-driver-user="root" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816627 4810 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816632 4810 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816636 4810 flags.go:64] FLAG: --system-cgroups="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816640 4810 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816668 4810 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816672 4810 flags.go:64] FLAG: --tls-cert-file="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816677 4810 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816682 4810 flags.go:64] FLAG: --tls-min-version="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816687 4810 flags.go:64] FLAG: --tls-private-key-file="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816691 4810 flags.go:64] FLAG: --topology-manager-policy="none" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816695 4810 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816700 4810 flags.go:64] FLAG: --topology-manager-scope="container" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816704 4810 flags.go:64] FLAG: --v="2" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816710 4810 flags.go:64] FLAG: --version="false" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816716 4810 flags.go:64] FLAG: --vmodule="" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816720 4810 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.816725 4810 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816850 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816854 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816859 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816863 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816867 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816871 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816877 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816880 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816884 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816887 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816893 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816896 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816900 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816903 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816907 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816911 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816914 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816918 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816922 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816925 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816929 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816932 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816936 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816939 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816943 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816947 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816952 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816956 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816981 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816985 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816990 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816994 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.816997 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817001 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817005 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817008 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817013 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817017 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817024 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817029 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817032 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817036 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817041 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817045 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817049 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817052 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817055 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817059 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817062 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817066 4810 feature_gate.go:330] unrecognized feature gate: Example Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817070 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817074 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817077 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817081 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817085 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817089 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817093 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817098 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817102 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817107 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817111 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817116 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817121 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817125 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817129 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817133 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817137 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817141 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817144 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817148 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.817153 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.817160 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.830903 4810 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.831030 4810 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831197 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831213 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831223 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831233 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831243 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831251 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831260 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831268 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831277 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831286 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831297 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831311 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831320 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831329 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831338 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831346 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831354 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831361 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831369 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831377 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831384 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831392 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831399 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831407 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831428 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831437 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831444 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831452 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831459 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831467 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831475 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831485 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831493 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831501 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831512 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831520 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831528 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831535 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831543 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831550 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831558 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831566 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831573 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831581 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831589 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831596 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831604 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831615 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831624 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831632 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831642 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831649 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831658 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831666 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831675 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831684 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831691 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831699 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831707 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831714 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831734 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831742 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831752 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831764 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831775 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831784 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831794 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831802 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831810 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831818 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.831826 4810 feature_gate.go:330] unrecognized feature gate: Example Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.831841 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832277 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832303 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832316 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832327 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832338 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832347 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832355 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832363 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832374 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832385 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832393 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832401 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832409 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832418 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832427 4810 feature_gate.go:330] unrecognized feature gate: Example Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832436 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832444 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832451 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832459 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832466 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832474 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832484 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832493 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832502 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832526 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832534 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832542 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832550 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832557 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832565 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832574 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832581 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832589 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832596 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832605 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832640 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832648 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832655 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832663 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832671 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832679 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832686 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832695 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832702 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832710 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832717 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832725 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832733 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832743 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832753 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832763 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832771 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832810 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832820 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832829 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832838 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832849 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832856 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832865 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832872 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832896 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832904 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832911 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832920 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832928 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832936 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832944 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832955 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.832991 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.833000 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.833008 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.833021 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.836274 4810 server.go:940] "Client rotation is on, will bootstrap in background" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.843183 4810 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.843346 4810 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.845377 4810 server.go:997] "Starting client certificate rotation" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.845430 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.845668 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-28 06:01:44.101168869 +0000 UTC Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.845796 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1943h30m6.255378548s for next certificate rotation Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.874738 4810 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.877740 4810 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.900989 4810 log.go:25] "Validated CRI v1 runtime API" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.938866 4810 log.go:25] "Validated CRI v1 image API" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.940911 4810 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.948921 4810 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-08-06-23-09-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.949022 4810 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.979938 4810 manager.go:217] Machine: {Timestamp:2025-10-08 06:31:37.975758921 +0000 UTC m=+0.610198741 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:dc4845f6-ce3e-4749-b720-9d8c65ff7503 BootID:e21113ec-53e0-4165-8cfb-266a2e1aef0b Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:a9:73:66 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:a9:73:66 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ba:ad:51 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:14:71:4e Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:8e:f3:95 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:74:9d:72 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:69:16:64 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:da:bc:78:5f:65:d7 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7a:cb:23:43:96:ff Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.980712 4810 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.980950 4810 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.981439 4810 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.981839 4810 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.981895 4810 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.982340 4810 topology_manager.go:138] "Creating topology manager with none policy" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.982363 4810 container_manager_linux.go:303] "Creating device plugin manager" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.983051 4810 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.983108 4810 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.984063 4810 state_mem.go:36] "Initialized new in-memory state store" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.984209 4810 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.988853 4810 kubelet.go:418] "Attempting to sync node with API server" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.988878 4810 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.988907 4810 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.988922 4810 kubelet.go:324] "Adding apiserver pod source" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.988936 4810 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.994240 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:37 crc kubenswrapper[4810]: E1008 06:31:37.994312 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:37 crc kubenswrapper[4810]: W1008 06:31:37.994478 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:37 crc kubenswrapper[4810]: E1008 06:31:37.994512 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.994526 4810 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.995703 4810 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 08 06:31:37 crc kubenswrapper[4810]: I1008 06:31:37.997478 4810 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000061 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000113 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000135 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000157 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000188 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000243 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000262 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000292 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000315 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000336 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000361 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000381 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.000445 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.001471 4810 server.go:1280] "Started kubelet" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.001716 4810 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.001738 4810 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.002327 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.002533 4810 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 08 06:31:38 crc systemd[1]: Started Kubernetes Kubelet. Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.005065 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.005123 4810 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.005236 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 21:43:25.165477282 +0000 UTC Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.005325 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1575h11m47.160162213s for next certificate rotation Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.005518 4810 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.005588 4810 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.005624 4810 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.005718 4810 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.007022 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="200ms" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.009059 4810 factory.go:55] Registering systemd factory Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.009104 4810 factory.go:221] Registration of the systemd container factory successfully Oct 08 06:31:38 crc kubenswrapper[4810]: W1008 06:31:38.009686 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.009786 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.009989 4810 factory.go:153] Registering CRI-O factory Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.010012 4810 factory.go:221] Registration of the crio container factory successfully Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.010068 4810 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.010087 4810 factory.go:103] Registering Raw factory Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.010099 4810 manager.go:1196] Started watching for new ooms in manager Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.011522 4810 manager.go:319] Starting recovery of all containers Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.013365 4810 server.go:460] "Adding debug handlers to kubelet server" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.011716 4810 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.102:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c705cbcc5f029 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 06:31:38.001412137 +0000 UTC m=+0.635851957,LastTimestamp:2025-10-08 06:31:38.001412137 +0000 UTC m=+0.635851957,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032219 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032385 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032449 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032536 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032615 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032692 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032849 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.032916 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033064 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033127 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033183 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033253 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033318 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033379 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033439 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033500 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033570 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033629 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033703 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033829 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033898 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.033974 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034036 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034093 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034146 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034201 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034268 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034327 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034387 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034444 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034500 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034555 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034614 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034670 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034725 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034791 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034849 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.034926 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035004 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035087 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035146 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035202 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035266 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035325 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035381 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035435 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035488 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035543 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035600 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035661 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035716 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035774 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035837 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035902 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.035983 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036090 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036180 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036240 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036304 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036361 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036424 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036482 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036544 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036599 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036658 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036714 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036768 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036820 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036878 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.036940 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037017 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037083 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037151 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037227 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037292 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037347 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037410 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037477 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037558 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037627 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037688 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037762 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037821 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037874 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.037927 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038004 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038062 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038117 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038173 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038242 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038310 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038371 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038435 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038490 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038546 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038612 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038683 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038745 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038806 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038861 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.038914 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039004 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039068 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039139 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039229 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039293 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039360 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039422 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039482 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039539 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039593 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039655 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039713 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039769 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039835 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039898 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.039956 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040044 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040124 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040198 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040283 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040349 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040409 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040467 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040533 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040589 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040653 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040761 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040819 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040875 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.040932 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041029 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041101 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041174 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041234 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041297 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041352 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041418 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041479 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041536 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041597 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041654 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041708 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041862 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.041927 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.042000 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.042060 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.042116 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.043813 4810 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.043903 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.043980 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044051 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044145 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044215 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044278 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044361 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044443 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044532 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044596 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044665 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044723 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044779 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044839 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044897 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.044959 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045052 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045121 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045180 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045237 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045316 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045384 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045443 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045504 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045564 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045642 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045722 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045783 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045837 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045895 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.045951 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046101 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046165 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046219 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046278 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046335 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046397 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046454 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046510 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046568 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046623 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046677 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046736 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046791 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046883 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.046944 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047018 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047085 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047142 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047196 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047248 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047300 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047362 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047419 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047471 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047528 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047581 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047635 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047697 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047766 4810 reconstruct.go:97] "Volume reconstruction finished" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.047819 4810 reconciler.go:26] "Reconciler: start to sync state" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.052514 4810 manager.go:324] Recovery completed Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.068451 4810 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.068750 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.071334 4810 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.071806 4810 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.071462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.071919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.071855 4810 kubelet.go:2335] "Starting kubelet main sync loop" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.071942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.072113 4810 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 08 06:31:38 crc kubenswrapper[4810]: W1008 06:31:38.072580 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.072642 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.076373 4810 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.076403 4810 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.076423 4810 state_mem.go:36] "Initialized new in-memory state store" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.097625 4810 policy_none.go:49] "None policy: Start" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.100023 4810 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.100058 4810 state_mem.go:35] "Initializing new in-memory state store" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.106386 4810 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.164452 4810 manager.go:334] "Starting Device Plugin manager" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.164565 4810 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.164592 4810 server.go:79] "Starting device plugin registration server" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.165525 4810 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.165569 4810 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.165750 4810 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.165909 4810 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.165926 4810 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.172626 4810 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.172801 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.174543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.174606 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.174625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.174865 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.175121 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.175182 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176108 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176198 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176646 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176876 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.176929 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.177836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.177876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.177881 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.177918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.177891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.177937 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.178180 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.178268 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.178304 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.178454 4810 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.178908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.178925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.178933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.179063 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.179410 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.179430 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.179660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.179700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.179718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.180099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.180158 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.180176 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.180498 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.180551 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.181143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.181172 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.181183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.181755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.181791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.181810 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.207621 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="400ms" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.249804 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.249872 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.249918 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.249953 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250175 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250237 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250420 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250510 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250566 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250612 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250659 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250716 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250767 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.250872 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.266277 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.268005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.268063 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.268080 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.268120 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.268573 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352437 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352537 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352664 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352725 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352796 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352840 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352931 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352932 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.352988 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353087 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353142 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353192 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353278 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353360 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353396 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353436 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353421 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353437 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353134 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353465 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.353561 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.469161 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.471403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.471484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.471513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.471558 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.472241 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.503921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.525118 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.553413 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: W1008 06:31:38.562333 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-90bce90b86b6c1e063dc2e1edccdc945bf8925b2eb727114ef63ac0ddd2a01c3 WatchSource:0}: Error finding container 90bce90b86b6c1e063dc2e1edccdc945bf8925b2eb727114ef63ac0ddd2a01c3: Status 404 returned error can't find the container with id 90bce90b86b6c1e063dc2e1edccdc945bf8925b2eb727114ef63ac0ddd2a01c3 Oct 08 06:31:38 crc kubenswrapper[4810]: W1008 06:31:38.566709 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-14fe5d22d7f23fad025fbab2fd121b903f67e298ebbe517582137da042837a6e WatchSource:0}: Error finding container 14fe5d22d7f23fad025fbab2fd121b903f67e298ebbe517582137da042837a6e: Status 404 returned error can't find the container with id 14fe5d22d7f23fad025fbab2fd121b903f67e298ebbe517582137da042837a6e Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.574666 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.583762 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:38 crc kubenswrapper[4810]: W1008 06:31:38.584912 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-4981ee9e3d23e759d758efe1c007fc5bf5072f2ee4f8c0e25e654d144f93795d WatchSource:0}: Error finding container 4981ee9e3d23e759d758efe1c007fc5bf5072f2ee4f8c0e25e654d144f93795d: Status 404 returned error can't find the container with id 4981ee9e3d23e759d758efe1c007fc5bf5072f2ee4f8c0e25e654d144f93795d Oct 08 06:31:38 crc kubenswrapper[4810]: W1008 06:31:38.606212 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-de499d4ad3e45ad61a855f6c292350e9eeb1b197b627ab2320e2f2cab6ad5aea WatchSource:0}: Error finding container de499d4ad3e45ad61a855f6c292350e9eeb1b197b627ab2320e2f2cab6ad5aea: Status 404 returned error can't find the container with id de499d4ad3e45ad61a855f6c292350e9eeb1b197b627ab2320e2f2cab6ad5aea Oct 08 06:31:38 crc kubenswrapper[4810]: W1008 06:31:38.607472 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-4d5e302fc3815458711c18859af841b39f1e20b345cae9af0d214b55503d7540 WatchSource:0}: Error finding container 4d5e302fc3815458711c18859af841b39f1e20b345cae9af0d214b55503d7540: Status 404 returned error can't find the container with id 4d5e302fc3815458711c18859af841b39f1e20b345cae9af0d214b55503d7540 Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.608943 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="800ms" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.872882 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.874492 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.874523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.874532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:38 crc kubenswrapper[4810]: I1008 06:31:38.874553 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 06:31:38 crc kubenswrapper[4810]: E1008 06:31:38.875506 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.003327 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:39 crc kubenswrapper[4810]: W1008 06:31:39.014482 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:39 crc kubenswrapper[4810]: E1008 06:31:39.014570 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:39 crc kubenswrapper[4810]: W1008 06:31:39.045321 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:39 crc kubenswrapper[4810]: E1008 06:31:39.045387 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.077388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"90bce90b86b6c1e063dc2e1edccdc945bf8925b2eb727114ef63ac0ddd2a01c3"} Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.082482 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4d5e302fc3815458711c18859af841b39f1e20b345cae9af0d214b55503d7540"} Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.083714 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"de499d4ad3e45ad61a855f6c292350e9eeb1b197b627ab2320e2f2cab6ad5aea"} Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.084567 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4981ee9e3d23e759d758efe1c007fc5bf5072f2ee4f8c0e25e654d144f93795d"} Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.085366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"14fe5d22d7f23fad025fbab2fd121b903f67e298ebbe517582137da042837a6e"} Oct 08 06:31:39 crc kubenswrapper[4810]: W1008 06:31:39.178547 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:39 crc kubenswrapper[4810]: E1008 06:31:39.178640 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:39 crc kubenswrapper[4810]: E1008 06:31:39.410462 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="1.6s" Oct 08 06:31:39 crc kubenswrapper[4810]: W1008 06:31:39.417381 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:39 crc kubenswrapper[4810]: E1008 06:31:39.417559 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:39 crc kubenswrapper[4810]: E1008 06:31:39.573862 4810 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.102:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c705cbcc5f029 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 06:31:38.001412137 +0000 UTC m=+0.635851957,LastTimestamp:2025-10-08 06:31:38.001412137 +0000 UTC m=+0.635851957,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.676118 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.677833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.677902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.677922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:39 crc kubenswrapper[4810]: I1008 06:31:39.678004 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 06:31:39 crc kubenswrapper[4810]: E1008 06:31:39.678701 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.003855 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.091012 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88" exitCode=0 Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.091136 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.091235 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.096531 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.096583 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.096599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.101795 4810 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851" exitCode=0 Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.101940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.102126 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.103631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.103658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.103673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.107594 4810 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d" exitCode=0 Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.107706 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.107884 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.109125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.109171 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.109188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.114134 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.114187 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.114211 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.114229 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.114193 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.115250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.115297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.115319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.117116 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135" exitCode=0 Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.117184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135"} Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.117403 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.118776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.118838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.118860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.123192 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.124304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.124360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:40 crc kubenswrapper[4810]: I1008 06:31:40.124377 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.004297 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:41 crc kubenswrapper[4810]: E1008 06:31:41.011296 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="3.2s" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.125252 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.125317 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.125327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.125437 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.132286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.132474 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.132561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.133734 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.133768 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.133781 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.135991 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa" exitCode=0 Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.136145 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.136155 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.137368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.137400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.137412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.139727 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.139806 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.140212 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"650cb4634edae47e869896bb3955e6ca75f5ddd7fadb02f36980c07b6bb7a3d6"} Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.143838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.143874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.143884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.143880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.144125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.144139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:41 crc kubenswrapper[4810]: W1008 06:31:41.189848 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:41 crc kubenswrapper[4810]: E1008 06:31:41.189937 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.279028 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.284147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.284191 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.284200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:41 crc kubenswrapper[4810]: I1008 06:31:41.284225 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 06:31:41 crc kubenswrapper[4810]: E1008 06:31:41.284694 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 06:31:41 crc kubenswrapper[4810]: W1008 06:31:41.619787 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 06:31:41 crc kubenswrapper[4810]: E1008 06:31:41.619861 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.145623 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e" exitCode=0 Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.145716 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e"} Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.145781 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.147264 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.147304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.147316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.152474 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.152485 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99"} Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.152531 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.152564 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5"} Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.152534 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.152633 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154161 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154201 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154889 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.154891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.707771 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.707958 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.710418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.710584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:42 crc kubenswrapper[4810]: I1008 06:31:42.710744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.158629 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f"} Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.158679 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.158685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53"} Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.158705 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073"} Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.158720 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.159538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.159580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.159592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.303835 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:43 crc kubenswrapper[4810]: I1008 06:31:43.713771 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.167007 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e"} Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.167068 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee"} Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.167105 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.167104 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.168368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.168405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.168420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.168485 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.168511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.168522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.485376 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.487152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.487207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.487224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.487266 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 06:31:44 crc kubenswrapper[4810]: I1008 06:31:44.985228 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.170190 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.170372 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.171956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.172073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.172099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.172185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.172251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.172272 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.914956 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.915311 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.917067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.917159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.917180 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.924187 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.977159 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.977410 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.978907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.979013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.979039 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:45 crc kubenswrapper[4810]: I1008 06:31:45.990038 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.172753 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.172801 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.172819 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.173142 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.174151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.174218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.174238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.175043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.175095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.175115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.175670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.175755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:46 crc kubenswrapper[4810]: I1008 06:31:46.175782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:47 crc kubenswrapper[4810]: I1008 06:31:47.174902 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:47 crc kubenswrapper[4810]: I1008 06:31:47.175885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:47 crc kubenswrapper[4810]: I1008 06:31:47.175942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:47 crc kubenswrapper[4810]: I1008 06:31:47.175954 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:47 crc kubenswrapper[4810]: I1008 06:31:47.866687 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:48 crc kubenswrapper[4810]: I1008 06:31:48.176429 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:48 crc kubenswrapper[4810]: I1008 06:31:48.177549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:48 crc kubenswrapper[4810]: I1008 06:31:48.177590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:48 crc kubenswrapper[4810]: I1008 06:31:48.177603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:48 crc kubenswrapper[4810]: E1008 06:31:48.179345 4810 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 06:31:49 crc kubenswrapper[4810]: I1008 06:31:49.572813 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:49 crc kubenswrapper[4810]: I1008 06:31:49.572946 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:49 crc kubenswrapper[4810]: I1008 06:31:49.574132 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:49 crc kubenswrapper[4810]: I1008 06:31:49.574362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:49 crc kubenswrapper[4810]: I1008 06:31:49.574509 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:50 crc kubenswrapper[4810]: I1008 06:31:50.867120 4810 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 06:31:50 crc kubenswrapper[4810]: I1008 06:31:50.867223 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 06:31:51 crc kubenswrapper[4810]: I1008 06:31:51.144654 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 08 06:31:51 crc kubenswrapper[4810]: I1008 06:31:51.145068 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:51 crc kubenswrapper[4810]: I1008 06:31:51.147046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:51 crc kubenswrapper[4810]: I1008 06:31:51.147120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:51 crc kubenswrapper[4810]: I1008 06:31:51.147163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.004234 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.066617 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.066682 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.072533 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.072797 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.187956 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.189438 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99" exitCode=255 Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.189505 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99"} Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.189663 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.190388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.190417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.190427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.190936 4810 scope.go:117] "RemoveContainer" containerID="9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99" Oct 08 06:31:52 crc kubenswrapper[4810]: I1008 06:31:52.230108 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:53 crc kubenswrapper[4810]: I1008 06:31:53.193430 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 06:31:53 crc kubenswrapper[4810]: I1008 06:31:53.195161 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174"} Oct 08 06:31:53 crc kubenswrapper[4810]: I1008 06:31:53.195315 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:53 crc kubenswrapper[4810]: I1008 06:31:53.196291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:53 crc kubenswrapper[4810]: I1008 06:31:53.196314 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:53 crc kubenswrapper[4810]: I1008 06:31:53.196321 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:53 crc kubenswrapper[4810]: I1008 06:31:53.304543 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:54 crc kubenswrapper[4810]: I1008 06:31:54.198012 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:54 crc kubenswrapper[4810]: I1008 06:31:54.199565 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:54 crc kubenswrapper[4810]: I1008 06:31:54.199636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:54 crc kubenswrapper[4810]: I1008 06:31:54.199658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:55 crc kubenswrapper[4810]: I1008 06:31:55.201648 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:55 crc kubenswrapper[4810]: I1008 06:31:55.203345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:55 crc kubenswrapper[4810]: I1008 06:31:55.203401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:55 crc kubenswrapper[4810]: I1008 06:31:55.203418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:55 crc kubenswrapper[4810]: I1008 06:31:55.998772 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:56 crc kubenswrapper[4810]: I1008 06:31:56.205316 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:31:56 crc kubenswrapper[4810]: I1008 06:31:56.206877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:31:56 crc kubenswrapper[4810]: I1008 06:31:56.207037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:31:56 crc kubenswrapper[4810]: I1008 06:31:56.207068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:31:56 crc kubenswrapper[4810]: I1008 06:31:56.212289 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:31:57 crc kubenswrapper[4810]: E1008 06:31:57.076037 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 08 06:31:57 crc kubenswrapper[4810]: E1008 06:31:57.091139 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.091394 4810 trace.go:236] Trace[1055101444]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 06:31:47.082) (total time: 10009ms): Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[1055101444]: ---"Objects listed" error: 10008ms (06:31:57.091) Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[1055101444]: [10.00900232s] [10.00900232s] END Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.091450 4810 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.092069 4810 trace.go:236] Trace[249027132]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 06:31:42.068) (total time: 15022ms): Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[249027132]: ---"Objects listed" error: 15022ms (06:31:57.091) Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[249027132]: [15.022938964s] [15.022938964s] END Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.092107 4810 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.093709 4810 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.095297 4810 trace.go:236] Trace[725916906]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 06:31:46.595) (total time: 10499ms): Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[725916906]: ---"Objects listed" error: 10499ms (06:31:57.094) Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[725916906]: [10.499350874s] [10.499350874s] END Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.095381 4810 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.099737 4810 trace.go:236] Trace[1156645960]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 06:31:42.163) (total time: 14935ms): Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[1156645960]: ---"Objects listed" error: 14934ms (06:31:57.098) Oct 08 06:31:57 crc kubenswrapper[4810]: Trace[1156645960]: [14.935630352s] [14.935630352s] END Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.099815 4810 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.874557 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:57 crc kubenswrapper[4810]: I1008 06:31:57.879655 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.001655 4810 apiserver.go:52] "Watching apiserver" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.005441 4810 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.006254 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.006928 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.007133 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.007260 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.009083 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.016322 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.020189 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.020364 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.022182 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.022801 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.022857 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.023273 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.023774 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.032022 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.033680 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.033692 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.034094 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.034274 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.036034 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.060203 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.083148 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.102126 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.107622 4810 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.123692 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.140162 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.155013 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.179136 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.196056 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201773 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201814 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201835 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201852 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201888 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201904 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201921 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201958 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.201997 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202023 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202047 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202074 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202101 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202125 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202174 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202201 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202226 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202248 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202273 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202298 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202345 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202366 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202402 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202420 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202448 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202481 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202516 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202535 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202552 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202571 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202608 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202632 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202649 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202666 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202686 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202723 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202742 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202759 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202805 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202829 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202852 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202878 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202903 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202929 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.202979 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203004 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203026 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203049 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203072 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203096 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203120 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203148 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203172 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203195 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203215 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203232 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203253 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203274 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203293 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203312 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203330 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203348 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203371 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203404 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203429 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203477 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203500 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203524 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203549 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203573 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203595 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203620 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203645 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203671 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203692 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203715 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203773 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203811 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203831 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203849 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203870 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203902 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203920 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203936 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.203955 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204002 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204021 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204038 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204054 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204069 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204090 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204107 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204123 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204139 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204157 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204177 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204302 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204322 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204340 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204360 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204378 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204397 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204416 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204434 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204453 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204471 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204487 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204506 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204522 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204539 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204557 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204574 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204607 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204624 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204645 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204665 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204683 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204724 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204744 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204767 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204788 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204808 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204847 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204869 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204888 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204910 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.204980 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205005 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205022 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205056 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205073 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205093 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205111 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205131 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205149 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205167 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205184 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205201 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205220 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205238 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205257 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205274 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205295 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205349 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205371 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205394 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205410 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205430 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205466 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205483 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205501 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205518 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205535 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205553 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205570 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205587 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205606 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205625 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.205662 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206102 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206369 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206406 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206428 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206448 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206577 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206598 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206618 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206640 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206662 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206683 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206704 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206794 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206818 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206859 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206879 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206901 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206920 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206944 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.206983 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207003 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207249 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207285 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207352 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207431 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207470 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207542 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207639 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207669 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207734 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.207764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.214718 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.214756 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.215075 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.215256 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.215754 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.216017 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.216188 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.216548 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.216621 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.216941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.217005 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.217306 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.217425 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.217627 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.218310 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.218707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.218866 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.219107 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.219152 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.219172 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.219141 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.219515 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.219688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.219948 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.220056 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.220182 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.220492 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.220693 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.220719 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.220671 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.220698 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.221057 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.221208 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.221067 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.221769 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.222172 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.222182 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.222393 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.222688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.222888 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.222481 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.223251 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.223444 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.223432 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.223581 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.223668 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.224106 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.224147 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.224258 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.224469 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:58.724431306 +0000 UTC m=+21.358871046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.224756 4810 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.224937 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.225448 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.225623 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.225728 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.225842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.225904 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.226096 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:58.726074248 +0000 UTC m=+21.360513988 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.226133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.226426 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.229162 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.229511 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.229498 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.229861 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230368 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230724 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230765 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230923 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231137 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231177 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231156 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230233 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231471 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231492 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231709 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.233314 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230796 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.234256 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230756 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.230731 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.234941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.235361 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.235178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.235626 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231240 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.231889 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.235694 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.236153 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.236201 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.236544 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.236927 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.237012 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.237464 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.237759 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.242100 4810 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.243205 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.243784 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.244255 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.246354 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.246648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.246830 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.247341 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.247584 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.247825 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.248175 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.248429 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.248716 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.249218 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.249645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.253417 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.254078 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.254327 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.254344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.254581 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.255146 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.255504 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.255722 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.255761 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.255779 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.255883 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:58.755851125 +0000 UTC m=+21.390290875 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.256249 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.255546 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.257140 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.257216 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.257708 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.257755 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.257867 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.257455 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.258876 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.259190 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.259614 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.260427 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.260483 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.261128 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.263021 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.263267 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.263477 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.263822 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.263893 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.263912 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.263931 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.264090 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:58.764064557 +0000 UTC m=+21.398504297 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.264112 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.264131 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.263735 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.264471 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.264536 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.264874 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.265036 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.265085 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.265825 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.266451 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.267219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.267263 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.266328 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.267734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.267775 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.268403 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.269266 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.269667 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.270860 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.271062 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.271531 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.271554 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.271577 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.271978 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.272751 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.274388 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.274510 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.274704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.275767 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.275809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.275862 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.276133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.276491 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.276552 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.276705 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.276749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.276868 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.277047 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.277225 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.277557 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.277798 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.277889 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.278133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.278322 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.278410 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.278416 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.278746 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.278785 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.279137 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.279266 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.279535 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.279690 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:31:58.779665258 +0000 UTC m=+21.414105008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.280024 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.280530 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.280813 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.281122 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.281710 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.282171 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.282464 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.282472 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.283850 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.286055 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.286782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.288211 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.288930 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.290356 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.290727 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.290934 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.291276 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.286260 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.293324 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.294856 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.296495 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.296647 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.308845 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309153 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309301 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309396 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309571 4810 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309658 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309739 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309811 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.309891 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310013 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310096 4810 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310173 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310250 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310344 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310430 4810 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310515 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310589 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310647 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310666 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310745 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310774 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310794 4810 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310813 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310832 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310848 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310867 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310887 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310905 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310920 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310938 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310954 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311000 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311014 4810 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311027 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311040 4810 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311053 4810 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311065 4810 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311077 4810 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311090 4810 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311102 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311114 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311129 4810 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311141 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311155 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311167 4810 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311179 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311194 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311206 4810 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.311218 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317383 4810 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317409 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317420 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317431 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317443 4810 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317458 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317468 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317478 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317490 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317500 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317510 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317520 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317530 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317540 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317553 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317565 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310071 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317579 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317711 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317723 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317734 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317745 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317756 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317766 4810 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317775 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317784 4810 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317795 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317804 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317814 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317826 4810 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317835 4810 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317845 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317906 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.310633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.317987 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318052 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318063 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318072 4810 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318103 4810 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318144 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318154 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318164 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318173 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318184 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318213 4810 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318224 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318234 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318244 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318253 4810 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318261 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318290 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318299 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318310 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318321 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318336 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318345 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318392 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318401 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318412 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318421 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318488 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318505 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318576 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318587 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318597 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318712 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318723 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.318732 4810 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319487 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319648 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319687 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319699 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319711 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319721 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319732 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319815 4810 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319848 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319860 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319870 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319879 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319888 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.319897 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320052 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320103 4810 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320119 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320135 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320159 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320183 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320197 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320213 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320231 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320244 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320258 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320272 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320285 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320299 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320312 4810 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320326 4810 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320340 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320355 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320369 4810 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320382 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320396 4810 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320411 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320444 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320458 4810 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320475 4810 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320489 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320510 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320527 4810 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320540 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320554 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320567 4810 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320580 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320593 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320606 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320619 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320632 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320646 4810 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320659 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320676 4810 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320689 4810 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320704 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320719 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320732 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320767 4810 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320783 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320799 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320812 4810 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320825 4810 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320838 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320851 4810 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320866 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320881 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320911 4810 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320925 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320939 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.320953 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321002 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321015 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321028 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321041 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321055 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321068 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321082 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321096 4810 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321112 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321126 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321149 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.321162 4810 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.328055 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.340632 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.344515 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.353853 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.359351 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.363145 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.374169 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.375094 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 06:31:58 crc kubenswrapper[4810]: W1008 06:31:58.375782 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-68df507dcae48cc9b81b268f507a8ba8b141aeabe919475c3954f88a5f21476f WatchSource:0}: Error finding container 68df507dcae48cc9b81b268f507a8ba8b141aeabe919475c3954f88a5f21476f: Status 404 returned error can't find the container with id 68df507dcae48cc9b81b268f507a8ba8b141aeabe919475c3954f88a5f21476f Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.424570 4810 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.424628 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.727863 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.728418 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.728103 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.728704 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:59.728680438 +0000 UTC m=+22.363120188 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.728620 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.729177 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:59.72916162 +0000 UTC m=+22.363601370 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.829776 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.829911 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830034 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:31:59.829953447 +0000 UTC m=+22.464393227 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:31:58 crc kubenswrapper[4810]: I1008 06:31:58.830089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830163 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830189 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830210 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830269 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830294 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830346 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830273 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:59.830256115 +0000 UTC m=+22.464695875 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:58 crc kubenswrapper[4810]: E1008 06:31:58.830437 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:31:59.830412929 +0000 UTC m=+22.464852709 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.234912 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a"} Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.234958 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"70061ad8769039508c9cf473c2e0c066e6870a36d8876cc4f2c7d9734cb7d278"} Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.239469 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4f06841ba8dee8233466b0b80a8d9e6bffd4cd3198b089d51754e73be23d7ee1"} Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.244807 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30"} Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.244836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0"} Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.244846 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"68df507dcae48cc9b81b268f507a8ba8b141aeabe919475c3954f88a5f21476f"} Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.263480 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.287217 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.303010 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.319197 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.340427 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.360823 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.381880 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.402756 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.422076 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.456032 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.476151 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.503429 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.524843 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.549801 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.577038 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.599176 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:31:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.738351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.738420 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.738539 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.738603 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:01.738585937 +0000 UTC m=+24.373025667 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.738637 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.738850 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:01.738808483 +0000 UTC m=+24.373248413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.839641 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.839867 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.839916 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:32:01.839868967 +0000 UTC m=+24.474308747 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:31:59 crc kubenswrapper[4810]: I1008 06:31:59.840033 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840182 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840227 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840228 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840256 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840269 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840292 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840363 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:01.84033837 +0000 UTC m=+24.474778140 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:31:59 crc kubenswrapper[4810]: E1008 06:31:59.840400 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:01.840385321 +0000 UTC m=+24.474825091 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.072469 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.072603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.072603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:00 crc kubenswrapper[4810]: E1008 06:32:00.072749 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:00 crc kubenswrapper[4810]: E1008 06:32:00.073272 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:00 crc kubenswrapper[4810]: E1008 06:32:00.073399 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.079355 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.080520 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.083052 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.084312 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.086550 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.087631 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.089045 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.091336 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.092784 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.094819 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.095930 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.097527 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.098711 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.099848 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.101082 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.102283 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.103698 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.104578 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.107275 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.108344 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.110210 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.111781 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.112769 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.114232 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.115219 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.116543 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.119162 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.119762 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.120927 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.121573 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.122613 4810 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.122757 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.124667 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.127159 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.128800 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.132216 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.133382 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.134715 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.135609 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.137127 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.137861 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.140075 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.141638 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.144534 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.145862 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.147601 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.149561 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.150831 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.151370 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.152417 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.152947 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.154071 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.154718 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 08 06:32:00 crc kubenswrapper[4810]: I1008 06:32:00.155252 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.178521 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.196120 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.197016 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.199315 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.220881 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.231429 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.246687 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.282031 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.305048 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.328670 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.380078 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.413623 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.477782 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.500184 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.517629 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.542619 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.553782 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.565116 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.576154 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.591109 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-gxpv2"] Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.591399 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.595537 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.595635 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.595906 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.596213 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.615512 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.627885 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.643082 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.660016 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/efd7c94c-a5a9-4cee-b74b-d6d50af1a41f-hosts-file\") pod \"node-resolver-gxpv2\" (UID: \"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\") " pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.660105 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgpfp\" (UniqueName: \"kubernetes.io/projected/efd7c94c-a5a9-4cee-b74b-d6d50af1a41f-kube-api-access-fgpfp\") pod \"node-resolver-gxpv2\" (UID: \"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\") " pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.662185 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.676848 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.701094 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.715619 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.729440 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.742230 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.760175 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:01Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.760486 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.760515 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.760541 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgpfp\" (UniqueName: \"kubernetes.io/projected/efd7c94c-a5a9-4cee-b74b-d6d50af1a41f-kube-api-access-fgpfp\") pod \"node-resolver-gxpv2\" (UID: \"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\") " pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.760561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/efd7c94c-a5a9-4cee-b74b-d6d50af1a41f-hosts-file\") pod \"node-resolver-gxpv2\" (UID: \"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\") " pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.760643 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/efd7c94c-a5a9-4cee-b74b-d6d50af1a41f-hosts-file\") pod \"node-resolver-gxpv2\" (UID: \"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\") " pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.760665 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.760718 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.760777 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:05.760760158 +0000 UTC m=+28.395199898 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.760809 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:05.760789279 +0000 UTC m=+28.395229019 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.779805 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgpfp\" (UniqueName: \"kubernetes.io/projected/efd7c94c-a5a9-4cee-b74b-d6d50af1a41f-kube-api-access-fgpfp\") pod \"node-resolver-gxpv2\" (UID: \"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\") " pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.861917 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.862044 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862163 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:32:05.862119979 +0000 UTC m=+28.496559719 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862183 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862247 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862267 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.862296 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862318 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:05.862309414 +0000 UTC m=+28.496749154 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862575 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862623 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862641 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:01 crc kubenswrapper[4810]: E1008 06:32:01.862791 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:05.862764106 +0000 UTC m=+28.497203846 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.906099 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-gxpv2" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.976886 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-fjrrk"] Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.977583 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.978713 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7vlk5"] Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.979606 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-txvr6"] Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.980756 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.982196 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-txvr6" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.983553 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.983945 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.984032 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.984143 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.984259 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.984504 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.986289 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.986513 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.986636 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.986763 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.986806 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 08 06:32:01 crc kubenswrapper[4810]: I1008 06:32:01.986977 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.028262 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.045747 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063695 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-kubelet\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063730 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2b5ec8eb-9a64-46ae-b881-535323e3b686-proxy-tls\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-system-cni-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063759 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9f654c73-c1a2-4292-abe0-830b46ed68f5-cni-binary-copy\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063779 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-socket-dir-parent\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063796 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-daemon-config\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063813 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-etc-kubernetes\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063828 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm4kr\" (UniqueName: \"kubernetes.io/projected/f331b5d8-b85f-4982-885c-9f4e265c2262-kube-api-access-hm4kr\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063851 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-os-release\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063945 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-system-cni-dir\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.063995 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-cnibin\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064021 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f331b5d8-b85f-4982-885c-9f4e265c2262-cni-binary-copy\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064046 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-hostroot\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064065 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2b5ec8eb-9a64-46ae-b881-535323e3b686-mcd-auth-proxy-config\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064083 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-netns\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064101 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-conf-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064114 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2b5ec8eb-9a64-46ae-b881-535323e3b686-rootfs\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064152 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-cni-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064181 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-cni-multus\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064228 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mclq\" (UniqueName: \"kubernetes.io/projected/9f654c73-c1a2-4292-abe0-830b46ed68f5-kube-api-access-5mclq\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064246 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064260 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f331b5d8-b85f-4982-885c-9f4e265c2262-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064390 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-cnibin\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064468 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-k8s-cni-cncf-io\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064498 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-multus-certs\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064529 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-os-release\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064567 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-cni-bin\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.064606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dngxd\" (UniqueName: \"kubernetes.io/projected/2b5ec8eb-9a64-46ae-b881-535323e3b686-kube-api-access-dngxd\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.071156 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.072334 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.072358 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.072380 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:02 crc kubenswrapper[4810]: E1008 06:32:02.072449 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:02 crc kubenswrapper[4810]: E1008 06:32:02.072626 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:02 crc kubenswrapper[4810]: E1008 06:32:02.072756 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.088759 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.130227 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.145718 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.162206 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.166743 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-cnibin\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.166892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-k8s-cni-cncf-io\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.166848 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-cnibin\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.166989 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-k8s-cni-cncf-io\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167017 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-multus-certs\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-os-release\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167089 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-multus-certs\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167120 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-cni-bin\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dngxd\" (UniqueName: \"kubernetes.io/projected/2b5ec8eb-9a64-46ae-b881-535323e3b686-kube-api-access-dngxd\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-os-release\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167467 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-cni-bin\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167530 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-kubelet\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167599 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2b5ec8eb-9a64-46ae-b881-535323e3b686-proxy-tls\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167645 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-kubelet\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167803 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-system-cni-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.167619 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-system-cni-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.168750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9f654c73-c1a2-4292-abe0-830b46ed68f5-cni-binary-copy\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.168774 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-socket-dir-parent\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.168812 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-daemon-config\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.168832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-etc-kubernetes\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.168895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-socket-dir-parent\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169083 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-etc-kubernetes\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169482 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9f654c73-c1a2-4292-abe0-830b46ed68f5-cni-binary-copy\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169558 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-daemon-config\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169619 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm4kr\" (UniqueName: \"kubernetes.io/projected/f331b5d8-b85f-4982-885c-9f4e265c2262-kube-api-access-hm4kr\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-os-release\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169664 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-system-cni-dir\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169685 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-cnibin\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169710 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f331b5d8-b85f-4982-885c-9f4e265c2262-cni-binary-copy\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169730 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-netns\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-hostroot\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169784 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2b5ec8eb-9a64-46ae-b881-535323e3b686-mcd-auth-proxy-config\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-conf-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169827 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2b5ec8eb-9a64-46ae-b881-535323e3b686-rootfs\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-cni-multus\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-cni-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169927 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mclq\" (UniqueName: \"kubernetes.io/projected/9f654c73-c1a2-4292-abe0-830b46ed68f5-kube-api-access-5mclq\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169946 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.169982 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f331b5d8-b85f-4982-885c-9f4e265c2262-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.170526 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f331b5d8-b85f-4982-885c-9f4e265c2262-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.170814 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-os-release\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.170852 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-system-cni-dir\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.170880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-cnibin\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.171095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2b5ec8eb-9a64-46ae-b881-535323e3b686-rootfs\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.171100 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-cni-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.171149 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-var-lib-cni-multus\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.171158 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-host-run-netns\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.171479 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-hostroot\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.171565 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f331b5d8-b85f-4982-885c-9f4e265c2262-cni-binary-copy\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.171612 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9f654c73-c1a2-4292-abe0-830b46ed68f5-multus-conf-dir\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.172137 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2b5ec8eb-9a64-46ae-b881-535323e3b686-mcd-auth-proxy-config\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.172917 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f331b5d8-b85f-4982-885c-9f4e265c2262-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.173463 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2b5ec8eb-9a64-46ae-b881-535323e3b686-proxy-tls\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.183874 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.192074 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mclq\" (UniqueName: \"kubernetes.io/projected/9f654c73-c1a2-4292-abe0-830b46ed68f5-kube-api-access-5mclq\") pod \"multus-txvr6\" (UID: \"9f654c73-c1a2-4292-abe0-830b46ed68f5\") " pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.194559 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dngxd\" (UniqueName: \"kubernetes.io/projected/2b5ec8eb-9a64-46ae-b881-535323e3b686-kube-api-access-dngxd\") pod \"machine-config-daemon-7vlk5\" (UID: \"2b5ec8eb-9a64-46ae-b881-535323e3b686\") " pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.195151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm4kr\" (UniqueName: \"kubernetes.io/projected/f331b5d8-b85f-4982-885c-9f4e265c2262-kube-api-access-hm4kr\") pod \"multus-additional-cni-plugins-fjrrk\" (UID: \"f331b5d8-b85f-4982-885c-9f4e265c2262\") " pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.202540 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.216717 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.248260 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.253971 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28"} Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.255431 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-gxpv2" event={"ID":"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f","Type":"ContainerStarted","Data":"cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e"} Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.255458 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-gxpv2" event={"ID":"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f","Type":"ContainerStarted","Data":"aee93769ff496a3532a7d7434f940e8b5f6afb08e263cf28d61fd12083d152ee"} Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.271329 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.296928 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.298698 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: W1008 06:32:02.307457 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf331b5d8_b85f_4982_885c_9f4e265c2262.slice/crio-f9953829c6115f65a9262f125ad1d2d718603824affc6ca042c7871b6f05382c WatchSource:0}: Error finding container f9953829c6115f65a9262f125ad1d2d718603824affc6ca042c7871b6f05382c: Status 404 returned error can't find the container with id f9953829c6115f65a9262f125ad1d2d718603824affc6ca042c7871b6f05382c Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.308648 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.314589 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-txvr6" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.326739 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: W1008 06:32:02.328347 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b5ec8eb_9a64_46ae_b881_535323e3b686.slice/crio-1572d3f29715271ff2f8a0e2513deb1c7591c9b90e5266235974b2d8036e97ad WatchSource:0}: Error finding container 1572d3f29715271ff2f8a0e2513deb1c7591c9b90e5266235974b2d8036e97ad: Status 404 returned error can't find the container with id 1572d3f29715271ff2f8a0e2513deb1c7591c9b90e5266235974b2d8036e97ad Oct 08 06:32:02 crc kubenswrapper[4810]: W1008 06:32:02.329440 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f654c73_c1a2_4292_abe0_830b46ed68f5.slice/crio-a594eab98ca6a0a65d85bcde0e2d4eed67a18f3f55381d8fd084162d7ab1cfe9 WatchSource:0}: Error finding container a594eab98ca6a0a65d85bcde0e2d4eed67a18f3f55381d8fd084162d7ab1cfe9: Status 404 returned error can't find the container with id a594eab98ca6a0a65d85bcde0e2d4eed67a18f3f55381d8fd084162d7ab1cfe9 Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.349873 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.368286 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.381618 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkrh9"] Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.382617 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.384874 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.385373 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.385404 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.385461 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.385568 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.387215 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.391820 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.396563 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.410413 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.432408 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.444145 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.466242 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473646 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-var-lib-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473677 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-bin\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-slash\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-etc-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473728 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-script-lib\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473742 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hzkl\" (UniqueName: \"kubernetes.io/projected/ee7b17fc-599e-4723-80fb-eb248dfe539d-kube-api-access-6hzkl\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473767 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-ovn\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473788 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovn-node-metrics-cert\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473809 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-env-overrides\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473826 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473841 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-node-log\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473853 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-config\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473871 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-kubelet\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473946 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-systemd-units\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.473995 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-ovn-kubernetes\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.474048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.474068 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-netns\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.474084 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-log-socket\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.474100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-netd\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.474122 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-systemd\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.478330 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.491225 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.504168 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.514325 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.532410 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.556730 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575265 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-env-overrides\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575336 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-node-log\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575358 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-config\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575375 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-kubelet\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-systemd-units\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575409 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-ovn-kubernetes\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575447 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575468 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-netd\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-netns\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-log-socket\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-systemd\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575543 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-slash\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575562 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-var-lib-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575579 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-bin\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575600 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hzkl\" (UniqueName: \"kubernetes.io/projected/ee7b17fc-599e-4723-80fb-eb248dfe539d-kube-api-access-6hzkl\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575628 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-etc-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575645 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-script-lib\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-ovn\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.575680 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovn-node-metrics-cert\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576589 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-netns\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576654 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-systemd-units\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-ovn-kubernetes\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-kubelet\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576738 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-netd\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576777 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576795 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-node-log\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576813 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-bin\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576844 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-systemd\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576713 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576894 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-var-lib-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.576872 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-slash\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.577012 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-log-socket\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.577077 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-ovn\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.577243 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-etc-openvswitch\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.577621 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-script-lib\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.577646 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-config\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.577698 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-env-overrides\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.582893 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovn-node-metrics-cert\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.585522 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.597760 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hzkl\" (UniqueName: \"kubernetes.io/projected/ee7b17fc-599e-4723-80fb-eb248dfe539d-kube-api-access-6hzkl\") pod \"ovnkube-node-qkrh9\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.602192 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.617011 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.633353 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.651794 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.664367 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.676629 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.690516 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.707262 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.712708 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.723720 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:02 crc kubenswrapper[4810]: W1008 06:32:02.726718 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee7b17fc_599e_4723_80fb_eb248dfe539d.slice/crio-ee0a6f39b0a4e372e72a213ffb6f6ad0a5266d45d7ea3f803f14dcf33a8b2539 WatchSource:0}: Error finding container ee0a6f39b0a4e372e72a213ffb6f6ad0a5266d45d7ea3f803f14dcf33a8b2539: Status 404 returned error can't find the container with id ee0a6f39b0a4e372e72a213ffb6f6ad0a5266d45d7ea3f803f14dcf33a8b2539 Oct 08 06:32:02 crc kubenswrapper[4810]: I1008 06:32:02.734258 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:02Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.262340 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41" exitCode=0 Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.262442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.262498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"ee0a6f39b0a4e372e72a213ffb6f6ad0a5266d45d7ea3f803f14dcf33a8b2539"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.265294 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.265339 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.265355 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"1572d3f29715271ff2f8a0e2513deb1c7591c9b90e5266235974b2d8036e97ad"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.268380 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerStarted","Data":"29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.268425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerStarted","Data":"a594eab98ca6a0a65d85bcde0e2d4eed67a18f3f55381d8fd084162d7ab1cfe9"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.271344 4810 generic.go:334] "Generic (PLEG): container finished" podID="f331b5d8-b85f-4982-885c-9f4e265c2262" containerID="ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a" exitCode=0 Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.271426 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerDied","Data":"ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.271471 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerStarted","Data":"f9953829c6115f65a9262f125ad1d2d718603824affc6ca042c7871b6f05382c"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.283488 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.300799 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.309523 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.320015 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.339467 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.358498 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.375094 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.395508 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.419569 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.433257 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.456542 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.489297 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.491950 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.507929 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.507984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.507995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.508134 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.526440 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.528611 4810 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.529003 4810 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.534125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.534159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.534168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.534185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.534195 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.543231 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.598818 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: E1008 06:32:03.610938 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.614726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.615108 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.615194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.615301 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.615381 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.625290 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: E1008 06:32:03.628907 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.632174 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.632230 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.632242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.632261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.632273 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.638746 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: E1008 06:32:03.648232 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.656041 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.656075 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.656086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.656099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.656120 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.659417 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: E1008 06:32:03.668369 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.670717 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.671891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.671913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.671922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.671933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.671941 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: E1008 06:32:03.683689 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: E1008 06:32:03.683810 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.685792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.685822 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.685832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.685844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.685855 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.687772 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.702589 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.716227 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.733018 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.748786 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.763416 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.777872 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.789738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.789813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.789834 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.789864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.789885 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.795476 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.815148 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.831285 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:03Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.893173 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.893454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.893524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.893593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.893661 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.996409 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.996443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.996452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.996467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:03 crc kubenswrapper[4810]: I1008 06:32:03.996476 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:03Z","lastTransitionTime":"2025-10-08T06:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.072998 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.073014 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.073156 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:04 crc kubenswrapper[4810]: E1008 06:32:04.073847 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:04 crc kubenswrapper[4810]: E1008 06:32:04.074483 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:04 crc kubenswrapper[4810]: E1008 06:32:04.074622 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.099000 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.099055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.099071 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.099093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.099107 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.201696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.201869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.201940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.202071 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.202156 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.279448 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerStarted","Data":"001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.283717 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.284002 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.284180 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.284341 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.284477 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.298220 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.304785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.305317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.305332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.305353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.305368 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.316512 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.337616 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.353894 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.367259 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.383570 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.408236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.408625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.408725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.408834 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.408927 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.414940 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.437661 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.450876 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.467708 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.481025 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.496163 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.512203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.512241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.512251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.512270 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.512281 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.513304 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.527910 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:04Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.615930 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.616024 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.616049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.616077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.616095 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.719882 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.719946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.719987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.720013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.720029 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.823021 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.823078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.823088 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.823113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.823128 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.926645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.926745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.926763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.926795 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:04 crc kubenswrapper[4810]: I1008 06:32:04.926816 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:04Z","lastTransitionTime":"2025-10-08T06:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.032427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.032552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.032583 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.032619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.032654 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.104786 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gcps7"] Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.106110 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.108423 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.108925 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.109339 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.111377 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.133190 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.136031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.136081 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.136097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.136121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.136137 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.156947 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.187149 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.206582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5k8w\" (UniqueName: \"kubernetes.io/projected/bd43533a-1d47-4662-8918-86d6381fada7-kube-api-access-j5k8w\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.206661 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43533a-1d47-4662-8918-86d6381fada7-host\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.206698 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bd43533a-1d47-4662-8918-86d6381fada7-serviceca\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.212368 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.234557 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.239078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.239139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.239156 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.239181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.239199 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.256959 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.274343 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.292694 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.295666 4810 generic.go:334] "Generic (PLEG): container finished" podID="f331b5d8-b85f-4982-885c-9f4e265c2262" containerID="001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6" exitCode=0 Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.295715 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerDied","Data":"001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.299774 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.308339 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5k8w\" (UniqueName: \"kubernetes.io/projected/bd43533a-1d47-4662-8918-86d6381fada7-kube-api-access-j5k8w\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.308401 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43533a-1d47-4662-8918-86d6381fada7-host\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.308441 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bd43533a-1d47-4662-8918-86d6381fada7-serviceca\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.309501 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43533a-1d47-4662-8918-86d6381fada7-host\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.310496 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bd43533a-1d47-4662-8918-86d6381fada7-serviceca\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.325091 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.344152 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.345504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.345551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.345563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.345583 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.345594 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.351032 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5k8w\" (UniqueName: \"kubernetes.io/projected/bd43533a-1d47-4662-8918-86d6381fada7-kube-api-access-j5k8w\") pod \"node-ca-gcps7\" (UID: \"bd43533a-1d47-4662-8918-86d6381fada7\") " pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.387536 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.406122 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.420695 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.430408 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gcps7" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.435171 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.449906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.449956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.449994 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.450023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.450039 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.468436 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.488010 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.508296 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.531571 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.548145 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.557187 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.557232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.557245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.557265 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.557282 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.567742 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.590584 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.610037 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.622952 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.642503 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.658360 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.660511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.660627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.660660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.660698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.660724 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.676251 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.702361 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.720205 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.735261 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.749630 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:05Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.763890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.763929 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.763941 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.764003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.764019 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.820083 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.820187 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.820350 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.820404 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.820543 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:13.820422674 +0000 UTC m=+36.454862434 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.820612 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:13.820574148 +0000 UTC m=+36.455013938 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.866701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.866754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.866765 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.866786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.866798 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.920828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.921041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921147 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:32:13.921119769 +0000 UTC m=+36.555559509 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921206 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921223 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.921224 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921236 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921288 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921302 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921311 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:13.921294243 +0000 UTC m=+36.555733983 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921312 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:05 crc kubenswrapper[4810]: E1008 06:32:05.921344 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:13.921335934 +0000 UTC m=+36.555775674 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.969519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.969557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.969565 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.969578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:05 crc kubenswrapper[4810]: I1008 06:32:05.969588 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:05Z","lastTransitionTime":"2025-10-08T06:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.072510 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.072636 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.072523 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:06 crc kubenswrapper[4810]: E1008 06:32:06.072802 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:06 crc kubenswrapper[4810]: E1008 06:32:06.072957 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:06 crc kubenswrapper[4810]: E1008 06:32:06.073189 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.073507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.073571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.073591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.073619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.073641 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.176858 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.176907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.176922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.176940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.176952 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.279106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.279168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.279185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.279208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.279226 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.303160 4810 generic.go:334] "Generic (PLEG): container finished" podID="f331b5d8-b85f-4982-885c-9f4e265c2262" containerID="05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177" exitCode=0 Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.303267 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerDied","Data":"05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.305718 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gcps7" event={"ID":"bd43533a-1d47-4662-8918-86d6381fada7","Type":"ContainerStarted","Data":"7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.305818 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gcps7" event={"ID":"bd43533a-1d47-4662-8918-86d6381fada7","Type":"ContainerStarted","Data":"9e3a3a0053f321607ca437cad0ea3b8d4a2bd00efb6aa07ddee7f32bc75fd37e"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.348316 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.368269 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.382757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.382846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.382865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.382897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.382916 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.395480 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.453688 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.476994 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.484541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.484575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.484584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.484600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.484609 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.498597 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.511500 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.529842 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.545709 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.557083 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.572048 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.586299 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.587273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.587309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.587321 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.587340 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.587356 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.605804 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.619076 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.632832 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.643118 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.659958 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.670575 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.680514 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.690229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.690299 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.690312 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.690352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.690364 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.703065 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.717833 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.733062 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.746699 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.764785 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.779062 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.791293 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.792953 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.793019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.793032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.793049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.793062 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.803687 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.813843 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.826859 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.838520 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:06Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.895670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.895712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.895720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.895736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.895746 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.999205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.999252 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.999271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.999290 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:06 crc kubenswrapper[4810]: I1008 06:32:06.999302 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:06Z","lastTransitionTime":"2025-10-08T06:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.101701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.101736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.101744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.101759 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.101771 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.211462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.211527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.211546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.211570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.211588 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.311593 4810 generic.go:334] "Generic (PLEG): container finished" podID="f331b5d8-b85f-4982-885c-9f4e265c2262" containerID="4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61" exitCode=0 Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.311684 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerDied","Data":"4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.313367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.313424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.313441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.313465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.313482 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.317443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.338813 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.360427 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.385560 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.401309 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.418262 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.418341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.418363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.418388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.418424 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.426211 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.445184 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.463783 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.478307 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.492845 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.505490 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.519105 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.522158 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.522193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.522271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.522692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.522760 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.531299 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.547596 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.561081 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.572307 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:07Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.625130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.625170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.625179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.625193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.625202 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.729452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.729495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.729507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.729525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.729538 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.832553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.832602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.832614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.832635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.832673 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.935927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.936012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.936035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.936063 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:07 crc kubenswrapper[4810]: I1008 06:32:07.936080 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:07Z","lastTransitionTime":"2025-10-08T06:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.039531 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.039599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.039617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.039643 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.039661 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.073849 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.073935 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:08 crc kubenswrapper[4810]: E1008 06:32:08.074060 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:08 crc kubenswrapper[4810]: E1008 06:32:08.074208 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.074350 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:08 crc kubenswrapper[4810]: E1008 06:32:08.074457 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.116148 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.137354 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.142805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.142861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.142879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.142903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.142921 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.157399 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.173283 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.206347 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.224348 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.240440 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.245625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.245681 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.245699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.245724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.245741 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.259352 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.282332 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.300408 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.316836 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.323370 4810 generic.go:334] "Generic (PLEG): container finished" podID="f331b5d8-b85f-4982-885c-9f4e265c2262" containerID="8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103" exitCode=0 Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.323406 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerDied","Data":"8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.335464 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: E1008 06:32:08.341365 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf331b5d8_b85f_4982_885c_9f4e265c2262.slice/crio-conmon-8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103.scope\": RecentStats: unable to find data in memory cache]" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.348175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.348210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.348225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.348245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.348259 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.354302 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.369742 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.386909 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.405826 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.422246 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.439483 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.452636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.452684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.452700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.452718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.452731 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.452904 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.466496 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.483412 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.494942 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.515879 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.532073 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.548519 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.556181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.556259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.556279 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.556310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.556332 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.580059 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.600566 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.644068 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.659762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.659819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.659835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.659857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.659871 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.666010 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.694041 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:08Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.765496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.765736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.765756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.765780 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.765798 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.869186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.869218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.869231 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.869247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.869260 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.973099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.973168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.973193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.973223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:08 crc kubenswrapper[4810]: I1008 06:32:08.973250 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:08Z","lastTransitionTime":"2025-10-08T06:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.076478 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.076542 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.076567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.076599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.076624 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.180405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.180497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.180514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.180534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.180548 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.283420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.283485 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.283503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.283533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.283552 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.332511 4810 generic.go:334] "Generic (PLEG): container finished" podID="f331b5d8-b85f-4982-885c-9f4e265c2262" containerID="bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c" exitCode=0 Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.332593 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerDied","Data":"bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.342680 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.343245 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.361164 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.389254 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.389638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.389689 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.389709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.389740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.389759 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.405931 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.417013 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.440453 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.456393 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.477200 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.492499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.492570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.492607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.492637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.492659 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.494065 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.506007 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.523256 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.537141 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.548565 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.574842 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.589625 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.597633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.597677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.597688 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.597708 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.597722 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.603134 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.622185 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.638314 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.653514 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.669163 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.684268 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.700601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.700632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.700640 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.700655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.700664 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.700885 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.714740 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.725205 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.734656 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.744070 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.759612 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.775279 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.799062 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.804549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.804598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.804616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.804645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.804669 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.812112 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.824048 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.839619 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:09Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.906701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.906750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.906763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.906787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:09 crc kubenswrapper[4810]: I1008 06:32:09.906803 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:09Z","lastTransitionTime":"2025-10-08T06:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.009596 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.009642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.009656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.009676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.009691 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.073091 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.073130 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:10 crc kubenswrapper[4810]: E1008 06:32:10.073272 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.073707 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:10 crc kubenswrapper[4810]: E1008 06:32:10.073810 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:10 crc kubenswrapper[4810]: E1008 06:32:10.073899 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.111857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.111905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.111916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.111932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.111944 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.214674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.214723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.214736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.214756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.214776 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.318119 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.318185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.318202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.318229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.318249 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.352437 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" event={"ID":"f331b5d8-b85f-4982-885c-9f4e265c2262","Type":"ContainerStarted","Data":"d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.352583 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.353528 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.379073 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.389040 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.403878 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.420331 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.420848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.420881 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.420893 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.420910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.420922 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.436330 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.452184 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.467234 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.486438 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.499247 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.520951 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.522588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.522639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.522656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.522675 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.522688 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.539849 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.556663 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.575928 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.594630 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.615675 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.628458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.628540 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.628558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.628585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.628604 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.652221 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.678202 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.700107 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.720710 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.730807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.730863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.730878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.730897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.730911 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.743269 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.759643 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.776004 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.790634 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.805408 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.824387 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.834207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.834253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.834264 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.834279 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.834291 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.839694 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.853955 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.885064 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.903909 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.918367 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.937016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.937069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.937084 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.937107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.937124 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:10Z","lastTransitionTime":"2025-10-08T06:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:10 crc kubenswrapper[4810]: I1008 06:32:10.942427 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:10Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.039349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.039386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.039397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.039414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.039426 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.141833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.141896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.141909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.141929 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.141942 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.247593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.247642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.247655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.247676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.247715 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.352543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.352605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.352616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.352638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.352830 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.356245 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.456238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.456296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.456310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.456326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.456336 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.559087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.559135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.559146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.559164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.559176 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.661821 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.661883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.661903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.661933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.661951 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.765934 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.766039 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.766062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.766091 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.766108 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.869613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.869671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.869684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.869704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.869719 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.974605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.974680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.974704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.974733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:11 crc kubenswrapper[4810]: I1008 06:32:11.974752 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:11Z","lastTransitionTime":"2025-10-08T06:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.072846 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.072916 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.073000 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:12 crc kubenswrapper[4810]: E1008 06:32:12.073093 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:12 crc kubenswrapper[4810]: E1008 06:32:12.073195 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:12 crc kubenswrapper[4810]: E1008 06:32:12.073443 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.077805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.077857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.077872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.077892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.077906 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.167848 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.181278 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.181378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.181404 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.181441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.181472 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.285578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.285650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.285669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.285696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.285716 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.365694 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/0.log" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.369898 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7" exitCode=1 Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.370043 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.371341 4810 scope.go:117] "RemoveContainer" containerID="d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.390957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.391030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.391046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.391070 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.391085 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.400906 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.425397 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.454563 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.480093 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.496324 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.496384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.496401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.496424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.496441 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.502772 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.519249 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.532235 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.554941 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.569544 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.583866 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.598932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.598987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.598999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.599016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.599029 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.604103 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.617487 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.629941 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.644114 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.661814 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:12Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.119621 6092 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120338 6092 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120598 6092 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.121408 6092 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 06:32:12.121441 6092 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 06:32:12.121507 6092 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 06:32:12.121519 6092 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 06:32:12.121528 6092 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 06:32:12.121575 6092 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 06:32:12.121588 6092 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 06:32:12.121605 6092 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 06:32:12.121619 6092 factory.go:656] Stopping watch factory\\\\nI1008 06:32:12.121621 6092 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 06:32:12.121646 6092 ovnkube.go:599] Stopped ovnkube\\\\nI1008 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:12Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.702735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.702767 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.702776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.702789 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.702798 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.805332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.805375 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.805384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.805398 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.805408 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.908403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.908445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.908456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.908470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:12 crc kubenswrapper[4810]: I1008 06:32:12.908480 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:12Z","lastTransitionTime":"2025-10-08T06:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.011025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.011072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.011087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.011107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.011119 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.114921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.114986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.114998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.115014 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.115024 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.218105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.218151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.218166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.218182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.218191 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.321152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.321199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.321211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.321228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.321244 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.375812 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/0.log" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.379500 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.379812 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.403575 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.420456 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.423329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.423387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.423397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.423414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.423424 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.436490 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.449800 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.465627 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.484804 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.516030 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.525618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.525669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.525678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.525693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.525703 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.533347 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.555258 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:12Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.119621 6092 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120338 6092 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120598 6092 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.121408 6092 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 06:32:12.121441 6092 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 06:32:12.121507 6092 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 06:32:12.121519 6092 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 06:32:12.121528 6092 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 06:32:12.121575 6092 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 06:32:12.121588 6092 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 06:32:12.121605 6092 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 06:32:12.121619 6092 factory.go:656] Stopping watch factory\\\\nI1008 06:32:12.121621 6092 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 06:32:12.121646 6092 ovnkube.go:599] Stopped ovnkube\\\\nI1008 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.576923 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.592654 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.611801 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.628139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.628207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.628228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.628259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.628281 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.632808 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.653851 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.668063 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.732426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.732483 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.732497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.732521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.732543 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.752271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.752328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.752348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.752373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.752393 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.769739 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.775353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.775407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.775417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.775439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.775450 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.795500 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.801497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.801542 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.801552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.801568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.801582 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.817717 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.827365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.827466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.827487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.827546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.827564 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.835733 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.835787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.835888 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.836004 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:29.835951803 +0000 UTC m=+52.470391543 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.835900 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.836213 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:29.836160799 +0000 UTC m=+52.470600579 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.851489 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.856991 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.857045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.857058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.857082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.857095 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.872703 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:13Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.873059 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.875706 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.875754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.875772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.875799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.875817 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.936762 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937017 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:32:29.936929336 +0000 UTC m=+52.571369086 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.937109 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.937197 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937398 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937460 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937485 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937398 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937550 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:29.937534781 +0000 UTC m=+52.571974731 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937555 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937573 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:13 crc kubenswrapper[4810]: E1008 06:32:13.937617 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:29.937604803 +0000 UTC m=+52.572044553 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.980251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.980330 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.980351 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.980381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:13 crc kubenswrapper[4810]: I1008 06:32:13.980405 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:13Z","lastTransitionTime":"2025-10-08T06:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.072414 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.072463 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.072464 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:14 crc kubenswrapper[4810]: E1008 06:32:14.072687 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:14 crc kubenswrapper[4810]: E1008 06:32:14.073065 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:14 crc kubenswrapper[4810]: E1008 06:32:14.073172 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.083393 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.083452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.083470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.083496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.083514 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.186059 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.186097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.186108 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.186124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.186139 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.288635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.288686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.288697 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.288712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.288722 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.385693 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/1.log" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.387331 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/0.log" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.390710 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49" exitCode=1 Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.390762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.390821 4810 scope.go:117] "RemoveContainer" containerID="d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.392252 4810 scope.go:117] "RemoveContainer" containerID="62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49" Oct 08 06:32:14 crc kubenswrapper[4810]: E1008 06:32:14.392642 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.393122 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.393183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.393207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.393240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.393267 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.431488 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:12Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.119621 6092 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120338 6092 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120598 6092 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.121408 6092 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 06:32:12.121441 6092 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 06:32:12.121507 6092 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 06:32:12.121519 6092 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 06:32:12.121528 6092 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 06:32:12.121575 6092 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 06:32:12.121588 6092 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 06:32:12.121605 6092 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 06:32:12.121619 6092 factory.go:656] Stopping watch factory\\\\nI1008 06:32:12.121621 6092 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 06:32:12.121646 6092 ovnkube.go:599] Stopped ovnkube\\\\nI1008 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.451529 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.473806 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.499643 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.500063 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.500165 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.500186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.500217 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.500245 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.522782 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.541212 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.559445 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.581635 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.597891 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.606531 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.606597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.606615 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.606639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.606659 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.624555 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.652618 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.685861 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.706852 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.711177 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.711220 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.711241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.711274 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.711295 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.727837 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.750614 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:14Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.813911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.814010 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.814076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.814099 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.814117 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.916247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.916668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.916813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.916914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:14 crc kubenswrapper[4810]: I1008 06:32:14.917034 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:14Z","lastTransitionTime":"2025-10-08T06:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.020444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.020833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.021037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.021201 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.021340 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.041289 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc"] Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.042567 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.048672 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.048860 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.050441 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a989c9a3-29fa-474b-8a2c-f1d1704d7623-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.050505 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a989c9a3-29fa-474b-8a2c-f1d1704d7623-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.050550 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a989c9a3-29fa-474b-8a2c-f1d1704d7623-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.050642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb7qm\" (UniqueName: \"kubernetes.io/projected/a989c9a3-29fa-474b-8a2c-f1d1704d7623-kube-api-access-bb7qm\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.065055 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.090358 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.109375 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.125767 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.126772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.126835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.126871 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.126898 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.126919 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.146023 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.151821 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb7qm\" (UniqueName: \"kubernetes.io/projected/a989c9a3-29fa-474b-8a2c-f1d1704d7623-kube-api-access-bb7qm\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.151892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a989c9a3-29fa-474b-8a2c-f1d1704d7623-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.151928 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a989c9a3-29fa-474b-8a2c-f1d1704d7623-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.152024 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a989c9a3-29fa-474b-8a2c-f1d1704d7623-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.153246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a989c9a3-29fa-474b-8a2c-f1d1704d7623-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.153429 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a989c9a3-29fa-474b-8a2c-f1d1704d7623-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.164067 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a989c9a3-29fa-474b-8a2c-f1d1704d7623-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.183019 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.192112 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb7qm\" (UniqueName: \"kubernetes.io/projected/a989c9a3-29fa-474b-8a2c-f1d1704d7623-kube-api-access-bb7qm\") pod \"ovnkube-control-plane-749d76644c-ktkpc\" (UID: \"a989c9a3-29fa-474b-8a2c-f1d1704d7623\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.206854 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.230007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.230075 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.230096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.230129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.230106 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.230155 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.265224 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.297158 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4604ca136850863daac83b3e04f81df1510116964b1706a7e403cd11cbf0da7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:12Z\\\",\\\"message\\\":\\\" from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.119621 6092 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120338 6092 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.120598 6092 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:12.121408 6092 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 06:32:12.121441 6092 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 06:32:12.121507 6092 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 06:32:12.121519 6092 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 06:32:12.121528 6092 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 06:32:12.121575 6092 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 06:32:12.121588 6092 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 06:32:12.121605 6092 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 06:32:12.121619 6092 factory.go:656] Stopping watch factory\\\\nI1008 06:32:12.121621 6092 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 06:32:12.121646 6092 ovnkube.go:599] Stopped ovnkube\\\\nI1008 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.317492 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.338353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.338440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.338465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.338502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.338529 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.344890 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.363663 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.373758 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: W1008 06:32:15.389478 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda989c9a3_29fa_474b_8a2c_f1d1704d7623.slice/crio-320be2d199457acc28250c671d32290cc6f295957e243f56c2282a8b38cf096d WatchSource:0}: Error finding container 320be2d199457acc28250c671d32290cc6f295957e243f56c2282a8b38cf096d: Status 404 returned error can't find the container with id 320be2d199457acc28250c671d32290cc6f295957e243f56c2282a8b38cf096d Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.410322 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.410810 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/1.log" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.419548 4810 scope.go:117] "RemoveContainer" containerID="62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49" Oct 08 06:32:15 crc kubenswrapper[4810]: E1008 06:32:15.419798 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.420246 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" event={"ID":"a989c9a3-29fa-474b-8a2c-f1d1704d7623","Type":"ContainerStarted","Data":"320be2d199457acc28250c671d32290cc6f295957e243f56c2282a8b38cf096d"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.440900 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.444365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.444412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.444423 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.444444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.444457 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.467716 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.504860 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.556170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.556236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.556249 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.556268 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.556282 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.561222 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.587446 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.608777 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.629238 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.644873 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.659400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.659440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.659450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.659466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.659478 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.663511 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.678437 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.695918 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.714653 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.732563 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.747439 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.762136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.762217 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.762241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.762272 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.762294 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.767379 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.781129 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.799387 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.817338 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:15Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.865269 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.865337 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.865358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.865385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.865405 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.968679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.968752 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.968772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.968797 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:15 crc kubenswrapper[4810]: I1008 06:32:15.968816 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:15Z","lastTransitionTime":"2025-10-08T06:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.072566 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:16 crc kubenswrapper[4810]: E1008 06:32:16.072704 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.072751 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.072831 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.072834 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.073007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.073043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: E1008 06:32:16.073082 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.073106 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.072790 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:16 crc kubenswrapper[4810]: E1008 06:32:16.073378 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.176019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.176080 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.176090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.176106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.176115 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.278850 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.278918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.278941 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.279021 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.279052 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.382517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.382589 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.382609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.382638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.382659 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.428087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" event={"ID":"a989c9a3-29fa-474b-8a2c-f1d1704d7623","Type":"ContainerStarted","Data":"80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.428174 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" event={"ID":"a989c9a3-29fa-474b-8a2c-f1d1704d7623","Type":"ContainerStarted","Data":"9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.466184 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.486001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.486042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.486053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.486068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.486078 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.489079 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.515492 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.538346 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.563214 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.577001 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.582064 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9s2m8"] Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.583107 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:16 crc kubenswrapper[4810]: E1008 06:32:16.583226 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.589630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.590145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.590157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.590188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.590201 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.595030 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.612860 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.626166 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.637919 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.654009 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.663621 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.667288 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c2gc\" (UniqueName: \"kubernetes.io/projected/669c3fc4-7586-4517-83c9-6316b2671329-kube-api-access-8c2gc\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.667488 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.680948 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.693526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.693577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.693624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.693651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.693666 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.694558 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.707812 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.720864 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.752995 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.768587 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.768663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c2gc\" (UniqueName: \"kubernetes.io/projected/669c3fc4-7586-4517-83c9-6316b2671329-kube-api-access-8c2gc\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:16 crc kubenswrapper[4810]: E1008 06:32:16.768779 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:16 crc kubenswrapper[4810]: E1008 06:32:16.768855 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:17.268834449 +0000 UTC m=+39.903274189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.774782 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.793936 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.796121 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c2gc\" (UniqueName: \"kubernetes.io/projected/669c3fc4-7586-4517-83c9-6316b2671329-kube-api-access-8c2gc\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.796228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.796265 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.796290 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.796311 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.796325 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.811594 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.832021 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.850253 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.871481 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.892804 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.898653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.898717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.898732 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.898752 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.898763 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:16Z","lastTransitionTime":"2025-10-08T06:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.912110 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.936455 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.952674 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:16 crc kubenswrapper[4810]: I1008 06:32:16.970272 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:16Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.002139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.002226 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.002247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.002279 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.002304 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.013459 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:17Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.037240 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:17Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.056826 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:17Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.075229 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:17Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.091462 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:17Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.106008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.106107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.106136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.106175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.106205 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.210089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.210154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.210167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.210188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.210202 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.274362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:17 crc kubenswrapper[4810]: E1008 06:32:17.274622 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:17 crc kubenswrapper[4810]: E1008 06:32:17.274789 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:18.274717283 +0000 UTC m=+40.909157033 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.314422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.314489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.314506 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.314530 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.314547 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.418805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.418891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.418915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.418950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.419020 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.523140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.523234 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.523246 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.523267 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.523278 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.626201 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.626276 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.626295 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.626319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.626339 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.729737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.729813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.729828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.729851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.729865 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.832544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.832608 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.832628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.832657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.832676 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.935602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.935677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.935698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.935731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:17 crc kubenswrapper[4810]: I1008 06:32:17.935752 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:17Z","lastTransitionTime":"2025-10-08T06:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.039418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.039474 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.039491 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.039517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.039535 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.072262 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.072328 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:18 crc kubenswrapper[4810]: E1008 06:32:18.072408 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.072534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:18 crc kubenswrapper[4810]: E1008 06:32:18.072580 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.072262 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:18 crc kubenswrapper[4810]: E1008 06:32:18.072738 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:18 crc kubenswrapper[4810]: E1008 06:32:18.072869 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.113706 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.137696 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.142533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.142602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.142615 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.142635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.142669 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.161698 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.188750 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.214606 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.237278 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.245157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.245232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.245255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.245282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.245299 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.258524 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.274934 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.286327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:18 crc kubenswrapper[4810]: E1008 06:32:18.286457 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:18 crc kubenswrapper[4810]: E1008 06:32:18.286528 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:20.286508701 +0000 UTC m=+42.920948441 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.287373 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.300616 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.312255 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.321760 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.334742 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.350184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.350233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.350245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.350263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.350278 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.356874 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.373787 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.401920 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.420083 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:18Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.452034 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.452078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.452090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.452106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.452119 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.555521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.555593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.555612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.556042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.556090 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.660164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.660212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.660228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.660247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.660263 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.764154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.764224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.764243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.764268 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.764284 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.867694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.867764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.867781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.867807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.867826 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.971756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.971817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.971837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.971864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:18 crc kubenswrapper[4810]: I1008 06:32:18.971884 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:18Z","lastTransitionTime":"2025-10-08T06:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.076404 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.076474 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.076494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.076522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.076543 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.180108 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.180203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.180224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.180252 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.180273 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.283547 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.283616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.283634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.283661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.283678 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.386798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.387317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.387468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.387833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.388041 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.498640 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.498722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.498748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.498786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.498813 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.602761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.602821 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.602843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.602870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.602887 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.705001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.705054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.705066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.705082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.705094 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.807925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.808000 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.808016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.808035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.808048 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.911899 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.912184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.912210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.912241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:19 crc kubenswrapper[4810]: I1008 06:32:19.912265 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:19Z","lastTransitionTime":"2025-10-08T06:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.015373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.015446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.015470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.015501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.015563 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.072291 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.072377 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.072395 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.072487 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:20 crc kubenswrapper[4810]: E1008 06:32:20.072479 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:20 crc kubenswrapper[4810]: E1008 06:32:20.072694 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:20 crc kubenswrapper[4810]: E1008 06:32:20.072823 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:20 crc kubenswrapper[4810]: E1008 06:32:20.073332 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.118641 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.118709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.118733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.118763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.118787 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.221814 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.221885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.221906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.221931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.221950 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.310475 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:20 crc kubenswrapper[4810]: E1008 06:32:20.310726 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:20 crc kubenswrapper[4810]: E1008 06:32:20.310898 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:24.310855869 +0000 UTC m=+46.945295659 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.324901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.324958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.325012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.325044 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.325066 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.428671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.428754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.428811 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.428844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.428870 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.531375 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.531450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.531473 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.531504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.531529 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.635283 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.635329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.635337 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.635353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.635362 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.738731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.738809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.738831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.738869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.738891 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.842369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.842418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.842440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.842470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.842491 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.945434 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.945508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.945527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.945551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:20 crc kubenswrapper[4810]: I1008 06:32:20.945571 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:20Z","lastTransitionTime":"2025-10-08T06:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.048334 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.048410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.048436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.048464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.048488 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.151954 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.152067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.152091 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.152128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.152149 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.254949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.255048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.255075 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.255106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.255128 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.359050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.359152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.359170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.359199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.359220 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.462876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.462928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.462948 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.463001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.463020 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.565594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.565663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.565682 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.565707 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.565726 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.668802 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.668879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.668902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.668934 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.668956 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.780585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.780638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.780655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.780679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.780697 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.883950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.884036 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.884052 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.884084 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.884106 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.987295 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.987365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.987391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.987423 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:21 crc kubenswrapper[4810]: I1008 06:32:21.987443 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:21Z","lastTransitionTime":"2025-10-08T06:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.072826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.073009 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:22 crc kubenswrapper[4810]: E1008 06:32:22.073080 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.073108 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.073104 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:22 crc kubenswrapper[4810]: E1008 06:32:22.073254 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:22 crc kubenswrapper[4810]: E1008 06:32:22.073584 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:22 crc kubenswrapper[4810]: E1008 06:32:22.075237 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.090694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.090744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.090762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.090786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.090804 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.193609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.193712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.193734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.193768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.193790 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.296860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.296951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.297011 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.297057 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.297082 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.399853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.399892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.399902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.399916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.399927 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.502868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.502940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.503002 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.503031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.503050 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.605851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.605922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.605934 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.605953 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.605987 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.708250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.708317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.708342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.708363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.708376 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.811078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.811119 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.811147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.811165 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.811175 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.914344 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.914414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.914457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.914490 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:22 crc kubenswrapper[4810]: I1008 06:32:22.914513 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:22Z","lastTransitionTime":"2025-10-08T06:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.017094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.017154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.017177 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.017206 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.017225 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.120419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.120521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.120541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.120563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.120616 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.223820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.223918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.223937 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.223988 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.224010 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.327098 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.327155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.327168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.327193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.327207 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.431331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.431413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.431431 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.431469 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.431490 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.534768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.534840 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.534857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.534883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.534901 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.638928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.639016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.639036 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.639060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.639078 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.742186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.742253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.742269 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.742296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.742313 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.845143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.845204 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.845223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.845247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.845264 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.948762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.948836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.948853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.948876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:23 crc kubenswrapper[4810]: I1008 06:32:23.948895 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:23Z","lastTransitionTime":"2025-10-08T06:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.051845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.051927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.051952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.052017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.052037 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.072795 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.072830 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.072819 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.073134 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.073261 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.073564 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.073657 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.073871 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.121633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.121693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.121710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.121734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.121753 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.143660 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:24Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.149000 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.149053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.149070 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.149095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.149112 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.170750 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:24Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.176672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.176730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.176750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.176810 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.176829 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.200736 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:24Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.206564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.206605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.206622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.206647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.206664 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.229666 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:24Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.234365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.234411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.234425 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.234446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.234458 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.251412 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:24Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.251641 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.253514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.253572 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.253587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.253608 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.253623 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.356808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.356911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.356935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.356989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.357009 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.359462 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.359625 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:24 crc kubenswrapper[4810]: E1008 06:32:24.359743 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:32.359709796 +0000 UTC m=+54.994149576 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.460302 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.460355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.460367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.460387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.460404 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.563479 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.563586 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.563607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.563633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.563652 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.667507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.667588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.667610 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.667644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.667666 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.777381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.777498 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.777519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.777550 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.777577 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.880953 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.881070 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.881089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.881114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.881131 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.984909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.985019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.985038 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.985066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:24 crc kubenswrapper[4810]: I1008 06:32:24.985086 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:24Z","lastTransitionTime":"2025-10-08T06:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.089025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.089130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.089169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.089211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.089242 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.195412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.195722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.195756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.195786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.195810 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.298787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.298872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.298891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.298922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.298945 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.402054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.402123 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.402146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.402175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.402197 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.505190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.505317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.505335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.505359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.505377 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.608213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.608270 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.608287 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.608310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.608327 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.711389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.711454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.711470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.711495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.711515 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.814533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.814598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.814622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.814650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.814673 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.917908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.917948 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.917958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.917997 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.918010 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:25Z","lastTransitionTime":"2025-10-08T06:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.984608 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 06:32:25 crc kubenswrapper[4810]: I1008 06:32:25.998147 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.008484 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.021260 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.021308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.021327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.021353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.021372 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.032684 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.052735 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.073498 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.073527 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.073640 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:26 crc kubenswrapper[4810]: E1008 06:32:26.073677 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.073742 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:26 crc kubenswrapper[4810]: E1008 06:32:26.073922 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:26 crc kubenswrapper[4810]: E1008 06:32:26.074253 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:26 crc kubenswrapper[4810]: E1008 06:32:26.074455 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.074744 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.094195 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.107751 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.123920 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.124517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.124600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.124620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.124649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.124677 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.139251 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.155029 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.166346 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.181639 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.214470 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.227065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.227115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.227124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.227140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.227151 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.234079 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.258354 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.278370 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.293240 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.326808 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:26Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.329625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.329811 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.329943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.330100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.330235 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.432553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.432845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.432956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.433092 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.433193 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.535613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.535679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.535701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.535735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.535756 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.638579 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.638664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.638685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.638716 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.638737 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.742846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.742905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.742923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.742948 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.743000 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.846359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.846425 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.846443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.846468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.846490 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.950191 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.950254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.950273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.950298 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:26 crc kubenswrapper[4810]: I1008 06:32:26.950314 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:26Z","lastTransitionTime":"2025-10-08T06:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.053918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.053999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.054016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.054039 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.054057 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.156834 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.156895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.156911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.156933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.156948 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.260375 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.260444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.260467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.260504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.260524 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.363927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.364026 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.364045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.364074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.364094 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.466501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.466557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.466571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.466593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.466606 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.570066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.570123 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.570140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.570162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.570179 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.673512 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.673578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.673603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.673629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.673648 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.776899 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.777026 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.777050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.777076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.777095 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.880452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.880515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.880538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.880561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.880574 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.983244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.983308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.983327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.983353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:27 crc kubenswrapper[4810]: I1008 06:32:27.983372 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:27Z","lastTransitionTime":"2025-10-08T06:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.072882 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.072888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.073047 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:28 crc kubenswrapper[4810]: E1008 06:32:28.073070 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.073161 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:28 crc kubenswrapper[4810]: E1008 06:32:28.073637 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:28 crc kubenswrapper[4810]: E1008 06:32:28.073790 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:28 crc kubenswrapper[4810]: E1008 06:32:28.073378 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.086472 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.086513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.086525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.086563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.086614 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.141485 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.164576 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.183374 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.187890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.187926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.187934 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.187949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.187980 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.199954 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.212052 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.239677 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.256048 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.276128 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.287726 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.290210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.290242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.290250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.290263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.290272 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.301532 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.313527 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.326517 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.339867 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.350474 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.359342 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.371561 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.381871 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.389846 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:28Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.392593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.392710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.392744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.392781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.392809 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.495751 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.495797 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.495808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.495824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.495835 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.597532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.597574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.597585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.597604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.597616 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.700027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.700143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.700161 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.700185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.700202 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.803587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.803646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.803663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.803688 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.803707 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.906867 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.906951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.907028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.907062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:28 crc kubenswrapper[4810]: I1008 06:32:28.907413 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:28Z","lastTransitionTime":"2025-10-08T06:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.015738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.015789 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.015802 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.015819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.015833 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.073952 4810 scope.go:117] "RemoveContainer" containerID="62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.118570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.118606 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.118615 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.118630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.118639 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.222120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.222608 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.222628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.222655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.222678 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.324791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.324829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.324842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.324858 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.324869 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.427662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.427704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.427716 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.427735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.427749 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.491380 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/1.log" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.497358 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.498254 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.517459 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.530390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.530420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.530428 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.530441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.530451 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.553342 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.566338 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.577069 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.590613 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.609812 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.622461 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.632346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.632389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.632400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.632415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.632427 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.635507 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.648018 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.660870 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.673506 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.683546 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.697884 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.708400 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.719855 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.729388 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.735164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.735204 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.735213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.735228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.735238 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.741820 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.751671 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:29Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.838041 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.838109 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.838128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.838153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.838172 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.927274 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.927381 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:29 crc kubenswrapper[4810]: E1008 06:32:29.927474 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:29 crc kubenswrapper[4810]: E1008 06:32:29.927548 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:29 crc kubenswrapper[4810]: E1008 06:32:29.927596 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:33:01.927567231 +0000 UTC m=+84.562006981 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:32:29 crc kubenswrapper[4810]: E1008 06:32:29.927630 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:33:01.927608952 +0000 UTC m=+84.562048732 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.940662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.940721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.940734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.940754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:29 crc kubenswrapper[4810]: I1008 06:32:29.940768 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:29Z","lastTransitionTime":"2025-10-08T06:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.028564 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.028662 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.028697 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.028743 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:33:02.028716527 +0000 UTC m=+84.663156257 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029150 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029185 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029200 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029266 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:33:02.02924885 +0000 UTC m=+84.663688590 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029726 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029759 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029774 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.029832 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:33:02.029820245 +0000 UTC m=+84.664259985 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.043286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.043331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.043341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.043357 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.043368 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.072908 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.072948 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.073052 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.073004 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.073228 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.073406 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.073448 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.073561 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.146665 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.146714 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.146726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.146745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.146758 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.250405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.250455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.250466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.250484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.250493 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.354212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.354277 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.354289 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.354315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.354330 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.456943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.457006 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.457020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.457037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.457049 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.505389 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/2.log" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.506859 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/1.log" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.512239 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee" exitCode=1 Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.512357 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.512627 4810 scope.go:117] "RemoveContainer" containerID="62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.513898 4810 scope.go:117] "RemoveContainer" containerID="ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee" Oct 08 06:32:30 crc kubenswrapper[4810]: E1008 06:32:30.514195 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.554147 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.559852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.559917 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.559936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.559992 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.560018 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.575405 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.594777 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.615820 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.634066 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.662664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.662774 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.662793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.662829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.662849 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.666340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b9d96c732e7e04ae5e64a1c687211dbbcc138ecc582f9cbcadbb038ad52f49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:13Z\\\",\\\"message\\\":\\\"vent on pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 06:32:13.251599 6219 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 3.49353ms)\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}\\\\nI1008 06:32:13.251649 6219 services_controller.go:360] Finished syncing service community-operators on namespace openshift-marketplace for network=default : 1.607472ms\\\\nI1008 06:32:13.251625 6219 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}\\\\nI1008 06:32:13.251677 6219 services_controller.go:360] Finished syncing service certified-operators on namespace openshift-marketplace for network=default : 3.010558ms\\\\nI1008 06:32:13.251681 6219 services_controller.go:356] Processing sync for service openshift-etcd/etcd for network=default\\\\nI1008 06:32:13.251690 6219 services_controller.go:356] Processing sync for service openshift-kube-storage-version-migrator-operator/metrics for network=default\\\\nF1008 06:32:13.251685 6219 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.689310 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.709890 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.726378 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.748339 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.765923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.766042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.766063 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.766093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.766117 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.767837 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.788689 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.812327 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.830431 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.853617 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.870414 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.871422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.871477 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.871495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.871524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.871545 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.886281 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.903452 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:30Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.980210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.980290 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.980315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.980346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:30 crc kubenswrapper[4810]: I1008 06:32:30.980372 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:30Z","lastTransitionTime":"2025-10-08T06:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.083902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.083993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.084023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.084051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.084074 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.187670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.187719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.187732 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.187751 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.187766 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.291240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.291307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.291324 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.291346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.291363 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.395037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.395091 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.395109 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.395129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.395145 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.497857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.497943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.498007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.498047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.498070 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.519527 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/2.log" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.524509 4810 scope.go:117] "RemoveContainer" containerID="ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee" Oct 08 06:32:31 crc kubenswrapper[4810]: E1008 06:32:31.524782 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.559301 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.580044 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.600452 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.602437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.602685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.602854 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.603060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.603239 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.623146 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.647414 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.666944 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.682518 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.702456 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.706616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.706666 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.706679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.706698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.706713 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.720872 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.740250 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.760954 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.784772 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.807524 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.810644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.811031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.811252 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.811429 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.811626 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.826565 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.854485 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.924245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.924287 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.924297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.924313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.924326 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:31Z","lastTransitionTime":"2025-10-08T06:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.928299 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.945404 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:31 crc kubenswrapper[4810]: I1008 06:32:31.962925 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:31Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.026497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.026545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.026557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.026575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.026587 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.073409 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.073448 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.073626 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:32 crc kubenswrapper[4810]: E1008 06:32:32.073846 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.074044 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:32 crc kubenswrapper[4810]: E1008 06:32:32.074050 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:32 crc kubenswrapper[4810]: E1008 06:32:32.074245 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:32 crc kubenswrapper[4810]: E1008 06:32:32.074362 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.130598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.130646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.130660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.130679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.130691 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.234289 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.234433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.234463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.234493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.234519 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.339087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.339138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.339153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.339179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.339193 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.442370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.442452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.442473 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.442503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.442523 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.455067 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:32 crc kubenswrapper[4810]: E1008 06:32:32.455281 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:32 crc kubenswrapper[4810]: E1008 06:32:32.455402 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:32:48.455369169 +0000 UTC m=+71.089808939 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.546788 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.546873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.546891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.547043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.547066 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.651512 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.651587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.651605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.651635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.651655 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.756255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.756329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.756354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.756391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.756414 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.860660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.860733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.860757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.860790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.860815 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.965300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.965447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.965471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.965505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:32 crc kubenswrapper[4810]: I1008 06:32:32.965524 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:32Z","lastTransitionTime":"2025-10-08T06:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.068870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.068916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.068933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.068960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.069006 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.173209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.173901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.173929 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.173959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.174042 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.278640 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.278723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.278744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.278779 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.278809 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.382591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.382661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.382680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.382719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.382748 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.485608 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.485704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.485721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.485745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.485765 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.588268 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.588331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.588348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.588378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.588396 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.691449 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.691497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.691515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.691538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.691559 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.796763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.796854 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.796873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.796904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.796925 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.900588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.900662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.900680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.900712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:33 crc kubenswrapper[4810]: I1008 06:32:33.900732 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:33Z","lastTransitionTime":"2025-10-08T06:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.004786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.004865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.004885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.004918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.004939 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.072774 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.072872 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.072797 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.072993 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.073082 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.073327 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.073563 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.073751 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.108469 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.108549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.108573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.108611 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.108634 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.212383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.212446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.212466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.212492 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.212510 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.295551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.295620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.295639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.295664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.295683 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.317421 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:34Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.325072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.325135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.325156 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.325206 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.325234 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.347497 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:34Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.358395 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.358502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.358532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.358577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.358615 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.382152 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:34Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.387561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.387617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.387633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.387712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.387732 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.405681 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:34Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.412085 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.412166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.412213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.412236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.412251 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.428774 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:34Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:34 crc kubenswrapper[4810]: E1008 06:32:34.429087 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.432450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.432493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.432504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.432521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.432536 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.534920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.535009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.535029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.535054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.535071 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.638207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.638282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.638307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.638353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.638381 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.741538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.741661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.741696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.741749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.741783 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.845045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.845186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.845212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.845236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.845256 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.948782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.949014 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.949041 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.949102 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:34 crc kubenswrapper[4810]: I1008 06:32:34.949122 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:34Z","lastTransitionTime":"2025-10-08T06:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.053147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.053228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.053248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.053284 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.053309 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.156900 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.157051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.157073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.157105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.157130 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.260549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.261046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.261221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.261450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.261591 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.365181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.365261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.365282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.365315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.365337 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.469603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.469683 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.469704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.469734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.469756 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.573188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.573246 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.573258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.573278 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.573293 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.677100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.677200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.677223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.677312 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.677337 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.781587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.781660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.781678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.781710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.781734 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.886554 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.886630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.886651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.886684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.886703 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.991014 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.991470 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.991696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.991902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:35 crc kubenswrapper[4810]: I1008 06:32:35.992197 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:35Z","lastTransitionTime":"2025-10-08T06:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.072544 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.072544 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.072694 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:36 crc kubenswrapper[4810]: E1008 06:32:36.073447 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:36 crc kubenswrapper[4810]: E1008 06:32:36.073496 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.072711 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:36 crc kubenswrapper[4810]: E1008 06:32:36.073570 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:36 crc kubenswrapper[4810]: E1008 06:32:36.073409 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.095376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.095458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.095475 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.095501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.095519 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.199754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.199823 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.199841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.199872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.199891 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.304418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.304499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.304522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.304588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.304616 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.408523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.408571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.408582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.408599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.408611 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.511741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.511790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.511800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.511816 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.511827 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.614754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.614839 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.614859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.614893 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.614915 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.718105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.718208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.718238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.718279 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.718310 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.821041 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.821116 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.821134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.821164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.821188 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.924844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.924907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.924927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.924953 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:36 crc kubenswrapper[4810]: I1008 06:32:36.925011 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:36Z","lastTransitionTime":"2025-10-08T06:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.028381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.028469 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.028496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.028532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.028555 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.132350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.132413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.132436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.132460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.132477 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.235585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.235655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.235674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.235700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.235719 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.338787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.338862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.338883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.338911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.338933 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.441837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.441891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.441903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.441928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.441941 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.545204 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.545258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.545270 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.545285 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.545324 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.648370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.648446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.648464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.648496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.648515 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.751012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.751067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.751077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.751093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.751106 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.853638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.853674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.853686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.853702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.853713 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.956068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.956128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.956136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.956157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:37 crc kubenswrapper[4810]: I1008 06:32:37.956169 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:37Z","lastTransitionTime":"2025-10-08T06:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.058074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.058115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.058127 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.058146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.058156 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.072276 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.072343 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:38 crc kubenswrapper[4810]: E1008 06:32:38.072440 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.072564 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:38 crc kubenswrapper[4810]: E1008 06:32:38.072693 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:38 crc kubenswrapper[4810]: E1008 06:32:38.072872 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.073118 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:38 crc kubenswrapper[4810]: E1008 06:32:38.073242 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.094601 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.113261 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.132187 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.157239 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.162839 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.162930 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.162949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.163391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.163645 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.178708 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.197896 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.234320 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.256037 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.266561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.266639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.266659 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.266690 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.266717 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.277410 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.300338 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.333123 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.357014 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.369604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.369639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.369647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.369661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.369669 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.382081 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.405409 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.429708 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.460680 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.472368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.472437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.472457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.472486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.472507 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.481372 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.505838 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:38Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.575803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.575873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.575884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.575908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.575920 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.679181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.679244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.679257 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.679275 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.679288 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.781828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.781885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.781899 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.781920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.781936 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.884658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.884692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.884702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.884714 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.884727 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.988262 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.988345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.988366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.988396 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:38 crc kubenswrapper[4810]: I1008 06:32:38.988416 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:38Z","lastTransitionTime":"2025-10-08T06:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.091426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.091546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.091569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.091601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.091620 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.195773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.195860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.195908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.195943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.195999 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.301018 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.301116 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.301140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.301176 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.301200 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.405627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.405699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.405723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.405757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.405779 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.509715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.509801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.509829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.509872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.509897 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.613225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.613325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.613343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.613367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.613385 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.716248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.716296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.716310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.716327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.716338 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.819791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.819862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.819880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.819907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.819925 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.922691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.922785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.922822 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.922860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:39 crc kubenswrapper[4810]: I1008 06:32:39.922902 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:39Z","lastTransitionTime":"2025-10-08T06:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.026857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.026909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.026920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.026939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.026951 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.073037 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.073077 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.073095 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.073152 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:40 crc kubenswrapper[4810]: E1008 06:32:40.073210 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:40 crc kubenswrapper[4810]: E1008 06:32:40.073310 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:40 crc kubenswrapper[4810]: E1008 06:32:40.073433 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:40 crc kubenswrapper[4810]: E1008 06:32:40.073532 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.133394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.133455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.133471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.133511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.133529 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.236426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.236481 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.236493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.236509 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.236522 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.340195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.340266 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.340280 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.340303 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.340318 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.443175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.443217 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.443228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.443244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.443256 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.545695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.545759 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.545785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.545818 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.545840 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.649879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.649938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.649956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.650012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.650031 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.753598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.753659 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.753673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.753693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.753707 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.858786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.858848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.858862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.858888 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.858902 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.962856 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.962959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.963020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.963051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:40 crc kubenswrapper[4810]: I1008 06:32:40.963101 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:40Z","lastTransitionTime":"2025-10-08T06:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.066539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.066605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.066661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.066695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.066718 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.170647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.170714 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.170738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.170772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.170793 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.274515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.274615 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.274633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.274663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.274686 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.378688 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.378764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.378785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.378818 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.378840 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.482317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.482403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.482421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.482492 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.482513 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.585637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.585781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.585805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.585867 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.585891 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.688856 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.688893 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.688905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.688924 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.688937 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.791528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.791592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.791609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.791633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.791651 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.895238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.895360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.895379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.895413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.895432 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.999097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.999185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.999210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.999246 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:41 crc kubenswrapper[4810]: I1008 06:32:41.999271 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:41Z","lastTransitionTime":"2025-10-08T06:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.072470 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:42 crc kubenswrapper[4810]: E1008 06:32:42.072723 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.073182 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.073278 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:42 crc kubenswrapper[4810]: E1008 06:32:42.073325 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.073388 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:42 crc kubenswrapper[4810]: E1008 06:32:42.073539 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:42 crc kubenswrapper[4810]: E1008 06:32:42.073658 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.102046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.102086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.102104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.102126 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.102147 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.205140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.205169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.205185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.205208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.205225 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.308249 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.308317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.308342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.308371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.308390 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.412112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.412162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.412177 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.412192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.412211 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.516090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.516163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.516189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.516211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.516223 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.619330 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.619373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.619382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.619402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.619417 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.722372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.722427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.722436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.722452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.722466 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.825957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.826066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.826077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.826095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.826105 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.930364 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.930462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.930487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.930521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:42 crc kubenswrapper[4810]: I1008 06:32:42.930545 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:42Z","lastTransitionTime":"2025-10-08T06:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.035260 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.035319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.035332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.035353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.035368 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.138539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.138604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.138616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.138637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.138652 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.241599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.241647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.241658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.241673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.241683 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.345040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.345069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.345076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.345089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.345098 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.448613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.448654 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.448669 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.448687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.448698 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.550889 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.550913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.550922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.550932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.550941 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.655422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.655462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.655471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.655488 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.655498 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.759388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.759464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.759490 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.759528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.759556 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.862781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.862857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.862876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.862906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.862927 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.965753 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.965845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.965878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.965926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:43 crc kubenswrapper[4810]: I1008 06:32:43.965958 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:43Z","lastTransitionTime":"2025-10-08T06:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.069469 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.069505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.069514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.069531 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.069542 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.073202 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.073319 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.073328 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.073577 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.073613 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.074376 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.074507 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.074610 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.075025 4810 scope.go:117] "RemoveContainer" containerID="ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee" Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.075610 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.171594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.171654 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.171672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.171696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.171714 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.274871 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.274925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.274943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.274996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.275015 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.378877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.378923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.378933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.378950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.378977 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.483346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.483425 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.483449 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.483477 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.483500 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.538369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.538402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.538411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.538428 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.538439 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.558647 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:44Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.565639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.565692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.565705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.565724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.565736 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.580734 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:44Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.586232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.586292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.586310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.586335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.586350 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.599409 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:44Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.602943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.603014 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.603032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.603051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.603064 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.622781 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:44Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.627786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.627809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.627820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.627838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.627852 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.646412 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:44Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:44 crc kubenswrapper[4810]: E1008 06:32:44.646776 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.649259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.649452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.649538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.649621 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.649720 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.752821 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.752895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.752914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.752948 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.753007 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.855617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.855674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.855685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.855703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.855712 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.959588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.959660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.959688 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.959719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:44 crc kubenswrapper[4810]: I1008 06:32:44.959739 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:44Z","lastTransitionTime":"2025-10-08T06:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.062803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.062878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.062902 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.062932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.062953 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.166633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.166705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.166731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.166767 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.166795 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.270356 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.270440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.270461 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.270496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.270519 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.374238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.374295 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.374309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.374331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.374346 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.477357 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.477407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.477418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.477444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.477456 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.585943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.586060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.586086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.586119 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.586144 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.689702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.690026 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.690169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.690301 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.690433 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.793450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.793539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.793558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.793578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.793595 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.896419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.896603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.896728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.896830 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.896923 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.999387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.999441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.999465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.999494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:45 crc kubenswrapper[4810]: I1008 06:32:45.999516 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:45Z","lastTransitionTime":"2025-10-08T06:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.072455 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.072625 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.072714 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.072794 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:46 crc kubenswrapper[4810]: E1008 06:32:46.073444 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:46 crc kubenswrapper[4810]: E1008 06:32:46.073577 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:46 crc kubenswrapper[4810]: E1008 06:32:46.073824 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:46 crc kubenswrapper[4810]: E1008 06:32:46.074092 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.103074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.103148 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.103168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.103197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.103216 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.206644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.206700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.206712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.206735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.206750 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.309879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.309940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.309957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.310008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.310024 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.414121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.414181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.414194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.414220 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.414238 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.517737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.517794 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.517805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.517823 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.518121 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.621800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.621853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.621861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.621876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.621886 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.724936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.725007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.725020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.725039 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.725051 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.828140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.828236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.828251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.828271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.828285 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.931759 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.931802 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.931815 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.931833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:46 crc kubenswrapper[4810]: I1008 06:32:46.931845 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:46Z","lastTransitionTime":"2025-10-08T06:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.034849 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.034891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.034899 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.034913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.034923 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.137303 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.137354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.137367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.137385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.137397 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.240140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.240183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.240192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.240205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.240215 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.343112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.343171 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.343190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.343215 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.343232 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.445680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.445728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.445738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.445754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.445769 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.547775 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.547833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.547851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.547875 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.547893 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.650139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.650170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.650179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.650192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.650200 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.753308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.753722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.753740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.753763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.753780 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.857055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.857151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.857177 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.857219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.857244 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.960264 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.960315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.960325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.960343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:47 crc kubenswrapper[4810]: I1008 06:32:47.960356 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:47Z","lastTransitionTime":"2025-10-08T06:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.062619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.062677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.062685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.062700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.062709 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.072151 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.072165 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.072203 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:48 crc kubenswrapper[4810]: E1008 06:32:48.072248 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.072263 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:48 crc kubenswrapper[4810]: E1008 06:32:48.072328 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:48 crc kubenswrapper[4810]: E1008 06:32:48.072399 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:48 crc kubenswrapper[4810]: E1008 06:32:48.072470 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.097111 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.109617 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.125784 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.138901 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.150288 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.165339 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.165814 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.165932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.165951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.165982 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.166357 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.175915 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.189052 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.202649 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.212664 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.220717 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.233699 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.242357 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.256568 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.267666 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.268609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.268632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.268640 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.268652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.268661 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.278896 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.296173 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.306760 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:48Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.370639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.370683 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.370692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.370706 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.370716 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.458376 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:48 crc kubenswrapper[4810]: E1008 06:32:48.458547 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:48 crc kubenswrapper[4810]: E1008 06:32:48.458623 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:33:20.458605892 +0000 UTC m=+103.093045632 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.472485 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.472521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.472528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.472545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.472554 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.574804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.574841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.574851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.574866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.574880 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.677372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.677405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.677415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.677429 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.677438 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.779647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.779693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.779703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.779720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.779732 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.882658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.882702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.882711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.882729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.882739 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.985908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.985960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.986001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.986023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:48 crc kubenswrapper[4810]: I1008 06:32:48.986035 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:48Z","lastTransitionTime":"2025-10-08T06:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.088552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.088593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.088602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.088618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.088628 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.192261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.192312 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.192325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.192342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.192360 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.295196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.295332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.295359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.295384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.295475 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.398812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.398847 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.398856 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.398870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.398879 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.502214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.502255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.502267 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.502284 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.502296 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.598567 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/0.log" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.598620 4810 generic.go:334] "Generic (PLEG): container finished" podID="9f654c73-c1a2-4292-abe0-830b46ed68f5" containerID="29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414" exitCode=1 Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.598648 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerDied","Data":"29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.599023 4810 scope.go:117] "RemoveContainer" containerID="29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.604241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.604288 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.604301 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.604317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.604329 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.622246 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.642840 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.661417 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.679643 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.693050 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.705572 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.707147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.707186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.707195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.707210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.707220 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.719136 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:48Z\\\",\\\"message\\\":\\\"2025-10-08T06:32:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0\\\\n2025-10-08T06:32:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0 to /host/opt/cni/bin/\\\\n2025-10-08T06:32:03Z [verbose] multus-daemon started\\\\n2025-10-08T06:32:03Z [verbose] Readiness Indicator file check\\\\n2025-10-08T06:32:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.733153 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.751091 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.765292 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.776830 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.787111 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.809609 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.810395 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.810431 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.810438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.810451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.810461 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.821622 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.833335 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.845392 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.857703 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.873767 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:49Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.913226 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.913275 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.913284 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.913298 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:49 crc kubenswrapper[4810]: I1008 06:32:49.913307 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:49Z","lastTransitionTime":"2025-10-08T06:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.015183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.015214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.015222 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.015237 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.015245 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.073427 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:50 crc kubenswrapper[4810]: E1008 06:32:50.073580 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.073924 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:50 crc kubenswrapper[4810]: E1008 06:32:50.074007 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.074052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:50 crc kubenswrapper[4810]: E1008 06:32:50.074091 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.074126 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:50 crc kubenswrapper[4810]: E1008 06:32:50.074159 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.117627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.117664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.117673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.117686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.117696 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.220259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.220312 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.220321 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.220336 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.220347 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.322630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.322686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.322699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.322743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.322758 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.425522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.425560 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.425572 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.425589 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.425603 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.533504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.533559 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.533574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.533590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.533603 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.604460 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/0.log" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.604534 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerStarted","Data":"2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.626069 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.635824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.635852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.635860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.635872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.635881 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.644707 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.658981 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.673477 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.685779 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.711507 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.724207 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.736475 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.738283 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.738337 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.738354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.738378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.738395 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.754197 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.770079 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.785773 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.801570 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.820181 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:48Z\\\",\\\"message\\\":\\\"2025-10-08T06:32:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0\\\\n2025-10-08T06:32:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0 to /host/opt/cni/bin/\\\\n2025-10-08T06:32:03Z [verbose] multus-daemon started\\\\n2025-10-08T06:32:03Z [verbose] Readiness Indicator file check\\\\n2025-10-08T06:32:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.829825 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.841588 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.841619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.841634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.841650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.841663 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.841764 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.853831 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.866370 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.879247 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:50Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.944186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.944241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.944260 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.944284 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:50 crc kubenswrapper[4810]: I1008 06:32:50.944302 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:50Z","lastTransitionTime":"2025-10-08T06:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.047578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.047614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.047624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.047637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.047646 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.149371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.149460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.149484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.149513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.149531 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.252048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.252098 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.252115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.252139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.252155 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.353818 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.353895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.353914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.353940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.354001 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.456216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.456259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.456275 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.456292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.456301 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.558536 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.558600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.558610 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.558625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.558635 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.661248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.661304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.661318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.661335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.661350 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.763687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.763746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.763766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.763790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.763809 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.866347 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.866414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.866431 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.866455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.866475 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.969840 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.969913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.969933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.969959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:51 crc kubenswrapper[4810]: I1008 06:32:51.970071 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:51Z","lastTransitionTime":"2025-10-08T06:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.086803 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.086938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:52 crc kubenswrapper[4810]: E1008 06:32:52.087064 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.086812 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:52 crc kubenswrapper[4810]: E1008 06:32:52.087187 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.087150 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:52 crc kubenswrapper[4810]: E1008 06:32:52.087470 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:52 crc kubenswrapper[4810]: E1008 06:32:52.087607 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.088839 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.088876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.088908 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.088931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.088950 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.192278 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.192335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.192355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.192384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.192407 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.295690 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.295945 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.296046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.296163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.296254 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.398914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.399178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.399247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.399319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.399397 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.502633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.502723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.502735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.502749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.502771 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.605637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.605702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.605723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.605786 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.605809 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.707830 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.707858 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.707866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.707880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.707890 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.810136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.811045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.811236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.811445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.811648 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.913683 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.914033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.914156 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.914260 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:52 crc kubenswrapper[4810]: I1008 06:32:52.914350 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:52Z","lastTransitionTime":"2025-10-08T06:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.017088 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.017151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.017163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.017181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.017196 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.120946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.121008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.121025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.121047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.121061 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.225756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.226168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.226288 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.226376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.226475 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.328794 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.328824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.328832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.328846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.328855 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.432522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.432594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.432614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.432639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.432657 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.536583 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.536644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.536663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.536689 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.536709 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.639585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.639655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.639677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.639707 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.639734 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.743447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.743510 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.743527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.743571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.743591 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.847463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.847532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.847550 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.847576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.847597 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.950644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.950749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.950806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.950831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:53 crc kubenswrapper[4810]: I1008 06:32:53.950847 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:53Z","lastTransitionTime":"2025-10-08T06:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.054124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.054186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.054202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.054225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.054244 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.072607 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.072653 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.072699 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.072792 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.072819 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.073049 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.073161 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.073354 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.157852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.157906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.157928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.157959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.158017 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.261696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.261775 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.261897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.261932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.261960 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.364673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.364733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.364745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.364762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.364774 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.468134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.468186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.468199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.468218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.468231 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.571241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.571300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.571322 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.571347 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.571363 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.674205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.674301 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.674318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.674339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.674383 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.680825 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.681042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.681130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.681159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.681179 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.702871 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:54Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.708432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.708498 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.708520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.708544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.708563 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.727156 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:54Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.732450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.732497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.732508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.732527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.732539 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.750166 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:54Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.754999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.755103 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.755172 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.755229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.755290 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.768425 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:54Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.773306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.773424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.773490 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.773557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.773628 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.810836 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e21113ec-53e0-4165-8cfb-266a2e1aef0b\\\",\\\"systemUUID\\\":\\\"dc4845f6-ce3e-4749-b720-9d8c65ff7503\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:54Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:54 crc kubenswrapper[4810]: E1008 06:32:54.811283 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.813180 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.813268 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.813337 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.813396 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.813456 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.916404 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.916451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.916462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.916483 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:54 crc kubenswrapper[4810]: I1008 06:32:54.916497 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:54Z","lastTransitionTime":"2025-10-08T06:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.019493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.019545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.019556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.019574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.019587 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.122464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.122511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.122521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.122535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.122545 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.225849 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.226396 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.226749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.227112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.227375 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.331629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.331729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.331764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.331880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.331915 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.435281 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.435345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.435362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.435385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.435403 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.538655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.538744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.538763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.538799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.538823 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.642193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.642763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.642887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.643042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.643144 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.745412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.745454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.745463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.745476 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.745488 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.848100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.848143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.848153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.848168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.848180 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.950850 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.951409 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.951493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.951570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:55 crc kubenswrapper[4810]: I1008 06:32:55.951633 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:55Z","lastTransitionTime":"2025-10-08T06:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.054167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.054495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.054645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.054789 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.054934 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.072869 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.072908 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:56 crc kubenswrapper[4810]: E1008 06:32:56.073101 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.073196 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.073249 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:56 crc kubenswrapper[4810]: E1008 06:32:56.073359 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:56 crc kubenswrapper[4810]: E1008 06:32:56.073456 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:56 crc kubenswrapper[4810]: E1008 06:32:56.073535 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.161922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.162032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.162057 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.162082 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.162106 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.266076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.266142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.266163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.266187 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.266203 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.369486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.369545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.369561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.369585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.369603 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.472394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.472475 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.472512 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.472543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.472568 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.575863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.575921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.575936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.575955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.575993 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.679012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.679096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.679120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.679152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.679177 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.781832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.781880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.781895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.781916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.781934 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.884933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.885028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.885047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.885071 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.885090 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.988940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.989033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.989053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.989080 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:56 crc kubenswrapper[4810]: I1008 06:32:56.989100 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:56Z","lastTransitionTime":"2025-10-08T06:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.074195 4810 scope.go:117] "RemoveContainer" containerID="ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.091645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.091687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.091698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.092540 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.092588 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.133577 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.195397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.195748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.195760 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.195876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.195892 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.298738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.298792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.298809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.298836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.298853 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.402367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.402423 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.402435 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.402454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.402467 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.505084 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.505143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.505157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.505175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.505188 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.608197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.608235 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.608243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.608258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.608268 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.634850 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/2.log" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.638049 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.638677 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.656579 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.672231 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.681763 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.709710 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.710463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.710493 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.710504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.710520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.710532 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.726082 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.749174 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.767507 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.787194 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.807543 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.812868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.813096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.813322 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.813343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.813356 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.827859 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.843146 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:48Z\\\",\\\"message\\\":\\\"2025-10-08T06:32:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0\\\\n2025-10-08T06:32:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0 to /host/opt/cni/bin/\\\\n2025-10-08T06:32:03Z [verbose] multus-daemon started\\\\n2025-10-08T06:32:03Z [verbose] Readiness Indicator file check\\\\n2025-10-08T06:32:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.860543 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.876867 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.899574 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.912134 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.916397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.916432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.916440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.916467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.916478 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:57Z","lastTransitionTime":"2025-10-08T06:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.928187 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.940069 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.953288 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94a873e-a682-4f0c-aa1e-2087b009fa91\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650cb4634edae47e869896bb3955e6ca75f5ddd7fadb02f36980c07b6bb7a3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:57 crc kubenswrapper[4810]: I1008 06:32:57.964133 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:57Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.019253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.019291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.019300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.019313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.019324 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.072312 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.072338 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.072393 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:32:58 crc kubenswrapper[4810]: E1008 06:32:58.072422 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.072457 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:32:58 crc kubenswrapper[4810]: E1008 06:32:58.072556 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:32:58 crc kubenswrapper[4810]: E1008 06:32:58.072623 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:32:58 crc kubenswrapper[4810]: E1008 06:32:58.072671 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.104271 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.122315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.122380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.122399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.122424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.122444 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.122934 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.144101 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.166483 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.182830 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:48Z\\\",\\\"message\\\":\\\"2025-10-08T06:32:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0\\\\n2025-10-08T06:32:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0 to /host/opt/cni/bin/\\\\n2025-10-08T06:32:03Z [verbose] multus-daemon started\\\\n2025-10-08T06:32:03Z [verbose] Readiness Indicator file check\\\\n2025-10-08T06:32:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.206876 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.225407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.225481 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.225502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.225532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.225557 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.228502 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.245831 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.260321 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.271848 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.286632 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.300235 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94a873e-a682-4f0c-aa1e-2087b009fa91\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650cb4634edae47e869896bb3955e6ca75f5ddd7fadb02f36980c07b6bb7a3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.311714 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.324266 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.329390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.329418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.329426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.329438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.329450 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.338643 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.351710 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.382734 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.400686 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.420200 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.433498 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.433569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.433586 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.433614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.433634 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.536659 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.536728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.536745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.536770 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.536786 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.640275 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.640329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.640341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.640361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.640375 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.644634 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/3.log" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.645445 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/2.log" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.649374 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" exitCode=1 Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.649436 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.649490 4810 scope.go:117] "RemoveContainer" containerID="ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.650816 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:32:58 crc kubenswrapper[4810]: E1008 06:32:58.651141 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.690230 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff23bf36b0791b61510bf36b490adf9f0a77dea455249cc5c1b2d4f69e855dee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:29Z\\\",\\\"message\\\":\\\"9.964707 6429 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.964821 6429 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.964906 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:29.965207 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 06:32:29.965368 6429 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1008 06:32:29.965401 6429 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:29.965752 6429 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 06:32:29.965810 6429 factory.go:656] Stopping watch factory\\\\nI1008 06:32:29.965825 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:58Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:57.986222 6792 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:57.985588 6792 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:57.986260 6792 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:57.992168 6792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 06:32:57.993247 6792 factory.go:656] Stopping watch factory\\\\nI1008 06:32:57.995557 6792 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1008 06:32:57.995575 6792 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1008 06:32:57.995621 6792 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:57.995645 6792 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 06:32:57.995710 6792 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.710865 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.734449 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:48Z\\\",\\\"message\\\":\\\"2025-10-08T06:32:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0\\\\n2025-10-08T06:32:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0 to /host/opt/cni/bin/\\\\n2025-10-08T06:32:03Z [verbose] multus-daemon started\\\\n2025-10-08T06:32:03Z [verbose] Readiness Indicator file check\\\\n2025-10-08T06:32:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.743748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.743789 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.743797 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.743812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.743825 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.753616 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.771813 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.790923 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.813104 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.831959 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.846484 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.846529 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.846540 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.846558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.846568 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.848050 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.863149 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94a873e-a682-4f0c-aa1e-2087b009fa91\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650cb4634edae47e869896bb3955e6ca75f5ddd7fadb02f36980c07b6bb7a3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.875369 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.899237 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.916417 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.931784 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.949033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.949069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.949078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.949093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.949101 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:58Z","lastTransitionTime":"2025-10-08T06:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.971577 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:58 crc kubenswrapper[4810]: I1008 06:32:58.986296 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.000220 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:58Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.018352 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.031795 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.052019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.052091 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.052130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.052339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.052373 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.156138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.156215 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.156261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.156286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.156304 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.259468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.259514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.259530 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.259551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.259569 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.363037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.363126 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.363157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.363188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.363211 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.466313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.466369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.466384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.466406 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.466424 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.569494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.569525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.569533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.569546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.569554 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.655275 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/3.log" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.661166 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:32:59 crc kubenswrapper[4810]: E1008 06:32:59.661685 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.672748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.672809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.672827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.672852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.672870 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.677387 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"669c3fc4-7586-4517-83c9-6316b2671329\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8c2gc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9s2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.706432 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18d2ecb-b8ea-4ddc-bf59-2a857af5e1ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b7c2fb44eb46ac5d201f9e771e0962b7cef680e1c6c371352a85a70a0993a53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59d8b00e0ffc28e96cb1efae44ca22661d0cd2d0bcfe18ea2987076502d5d33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a421af723372ae5d314a52f0eec12dcbf50c3bb4d07b68cb5b5833eb73b9c8ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6fcbe2057e574c36793e885007671574c273a8e22e5e5203f75c2d9cf2e613e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe070f60e3bd3c3058833c772d15c3c206932b8aa7302d117f5bd2e7211a073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e69b29302085f08953a9cc7feedd8546da9ffc9126cb181a84199283d76b5b88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d947ed4355a58d1770bc83f768494f064949fb82c2db13ca1e86020837c168fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f13e13a2a9bbf8b832376bf1c5d2a025ef2b39f554249a4ace2c4ce0e565c99e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.723674 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.740372 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c32deb3b3e334b40c8e181a0617a7480062a0a75171c5fcdf65ec588f61a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.757782 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.775772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.775842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.775860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.775885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.775904 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.779439 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee7b17fc-599e-4723-80fb-eb248dfe539d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:58Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:57.986222 6792 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:57.985588 6792 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 06:32:57.986260 6792 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 06:32:57.992168 6792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 06:32:57.993247 6792 factory.go:656] Stopping watch factory\\\\nI1008 06:32:57.995557 6792 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1008 06:32:57.995575 6792 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1008 06:32:57.995621 6792 ovnkube.go:599] Stopped ovnkube\\\\nI1008 06:32:57.995645 6792 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 06:32:57.995710 6792 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hzkl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qkrh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.793705 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.810717 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b269f905820dcb0939cb90865050f08f3443da5d351b88722a948847cf182a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d77b6c9a6fb1fb0c78aa30461dedf285e5a256332bdf48a0a91ae604611d96c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.828318 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-txvr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f654c73-c1a2-4292-abe0-830b46ed68f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T06:32:48Z\\\",\\\"message\\\":\\\"2025-10-08T06:32:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0\\\\n2025-10-08T06:32:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c78ad136-8675-4471-9fdf-a9ff761629e0 to /host/opt/cni/bin/\\\\n2025-10-08T06:32:03Z [verbose] multus-daemon started\\\\n2025-10-08T06:32:03Z [verbose] Readiness Indicator file check\\\\n2025-10-08T06:32:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5mclq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-txvr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.847340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23ab48f1-b686-47ad-a18b-f449ff710511\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c4af0c78615126fe06c05dce5f3812ad5ed79f7985563f3faaf83afad331a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dacf6a39f22838fafc926db475daaebd442cc56e1a8ea3d9916b08de64a56db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f90d919c495b8b90aa81f2823afa8c308217cf735b494b9c6d3f34452398d8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fc18af51aacef73095739a06dcec44892373f203e724bb4e247c8244ff73174\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a3ef42ba95b1be4cf806cf531fb44e34d8d35f543c5d4bf5396bad0e0283c99\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T06:31:51Z\\\",\\\"message\\\":\\\"W1008 06:31:41.430722 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 06:31:41.431154 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759905101 cert, and key in /tmp/serving-cert-750304950/serving-signer.crt, /tmp/serving-cert-750304950/serving-signer.key\\\\nI1008 06:31:41.646034 1 observer_polling.go:159] Starting file observer\\\\nW1008 06:31:41.649221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 06:31:41.649447 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 06:31:41.650326 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-750304950/tls.crt::/tmp/serving-cert-750304950/tls.key\\\\\\\"\\\\nF1008 06:31:51.952152 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b40c2fd9c9d1402a96bf71485d3dbda1bf665c4959a0174698471f92f47df5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b52a2aac680aed33b88e4632bd12f7535444ad66a87d5453209ade044b32135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.869785 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7eec93c-d70b-43a9-8c29-163a062b6fa6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8a4d7bab641e59671492dfb13eecf7be089e0fd87d50bda4cb9303b2b2c3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c40d6a0c38fb58930a674f4d1369b71d0376b5468d5f5b7f6da38d6cb7a2744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99be3758147bec27c99c0eba013feab13b43778a377d8fae85d14c05b54f2b01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9ddf3f9d0893ad98003b93625bed9a3bbb58f1d3f46ffcc6b10140865fcd69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.879147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.879184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.879194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.879213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.879224 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.888433 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dcc280d-9985-4d0c-9e74-d1433d3b60f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e189005e176ac25bb4238190ec0a24efabe8e78203755a660c94306d36436f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23489e2b5e6eeb6a17a325e080da6a9729cbac00d3f28d9b10c60441c392466c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bbea1a0e5f5f60ac07e45d6abccde232e29fcb4297c2e56875b336234a7c7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0275fcf96f95c607bfbdd4182380467e16a2a199b723ee70129d3311cde36b2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.908955 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c85e2849cce27edf7dfef4c1cc6beff5cdac2086ee6d9c512c8a98cbf435c15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.927168 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gcps7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd43533a-1d47-4662-8918-86d6381fada7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7197d75eefd351d3e66ea14fa31f6063dec2022a87eec178350ba639c2c8c28b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5k8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gcps7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.943503 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a989c9a3-29fa-474b-8a2c-f1d1704d7623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9468b60cbe3f20b46fecbb220121574e204f2c0b3dd9cd08f6f82c518d4abd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80ee83da6c33221953cf7688c65cabc6e8ebfca17856b9b4fdc429acd94b3ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bb7qm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ktkpc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.963895 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94a873e-a682-4f0c-aa1e-2087b009fa91\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:31:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650cb4634edae47e869896bb3955e6ca75f5ddd7fadb02f36980c07b6bb7a3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b287e3fd73b8e948aad578ddecc6fcdb52d8982102d895821fdeb583a4d4851\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:31:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:31:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:31:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.979933 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-gxpv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efd7c94c-a5a9-4cee-b74b-d6d50af1a41f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc2a4fc4f5c9cf638952970735c2611952c651c29fedb5995947094acb8d358e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fgpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-gxpv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:32:59Z is after 2025-08-24T17:21:41Z" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.981819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.981887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.981905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.981930 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:32:59 crc kubenswrapper[4810]: I1008 06:32:59.981949 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:32:59Z","lastTransitionTime":"2025-10-08T06:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.006280 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f331b5d8-b85f-4982-885c-9f4e265c2262\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9c96291780ffa07d373e459a05f3a564e487ddaf041e776f56c67987b0c1321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceb016462b6e9b961c3b5b02dd5be662a13a6f5b3c513353bf13bbccb8a0d41a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://001db0f836252a9b8dc6ce1ba2d8400e3e5a18e0fcc85862a803fb3d0af79bc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c4c2b9905071d2759f95849ec05e5a3a02e5ea6d2dd4fa330eff039450b177\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b2396c0b91ed4ada46eed87f44c8b65c026d46820da5466f36ac117cc245d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8efd918cc3b3aa9e064dbcf6b01efcf206afefa591854bc9d14cffc81c9cd103\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcc67e67dcefed5ffd59a1a714211c0695ba37fe21db28387caa13c7c6b5d98c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T06:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T06:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hm4kr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fjrrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:33:00Z is after 2025-08-24T17:21:41Z" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.024312 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b5ec8eb-9a64-46ae-b881-535323e3b686\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T06:32:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3f60217091c1af383e9747a70ebe0e4175f66e4b38e4c60ebaa36ea693966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T06:32:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dngxd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T06:32:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7vlk5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T06:33:00Z is after 2025-08-24T17:21:41Z" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.073161 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:00 crc kubenswrapper[4810]: E1008 06:33:00.073374 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.073732 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:00 crc kubenswrapper[4810]: E1008 06:33:00.073861 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.074094 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:00 crc kubenswrapper[4810]: E1008 06:33:00.074210 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.074412 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:00 crc kubenswrapper[4810]: E1008 06:33:00.074555 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.084808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.084896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.084912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.084932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.084949 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.188345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.188405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.188427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.188454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.188476 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.291496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.291571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.291595 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.291620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.291638 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.394904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.394995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.395013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.395036 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.395053 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.498193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.498349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.498378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.498408 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.498430 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.601475 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.601553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.601568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.601589 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.601601 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.704742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.704812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.704827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.704887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.704906 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.808561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.808614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.808632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.808657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.808685 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.912280 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.912366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.912390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.912419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:00 crc kubenswrapper[4810]: I1008 06:33:00.912443 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:00Z","lastTransitionTime":"2025-10-08T06:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.016327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.016387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.016402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.016425 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.016443 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.119646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.119714 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.119730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.119754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.119769 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.222550 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.222663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.222682 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.222749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.222769 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.325849 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.326012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.326066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.326093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.326110 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.430147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.430200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.430211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.430230 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.430245 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.533418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.533487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.533505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.533529 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.533547 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.636052 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.636087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.636098 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.636115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.636126 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.738721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.738762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.738773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.738790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.738801 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.841721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.841788 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.841807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.841832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.841849 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.944879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.944955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.945004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.945030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:01 crc kubenswrapper[4810]: I1008 06:33:01.945047 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:01Z","lastTransitionTime":"2025-10-08T06:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.001033 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.001135 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.001261 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.001337 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.001381 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.00135493 +0000 UTC m=+148.635794670 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.001414 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.001393761 +0000 UTC m=+148.635833531 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.048347 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.048413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.048438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.048468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.048488 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.072609 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.072674 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.072860 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.072943 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.073130 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.073224 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.073334 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.073441 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.102441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.102627 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.102603741 +0000 UTC m=+148.737043501 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.102800 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.102857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103120 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103155 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103174 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103193 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103201 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103217 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103287 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.103267398 +0000 UTC m=+148.737707178 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:33:02 crc kubenswrapper[4810]: E1008 06:33:02.103316 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.103302999 +0000 UTC m=+148.737742779 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.151580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.151650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.151668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.151693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.151711 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.254562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.254634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.254661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.254690 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.254711 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.357724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.357828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.357844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.357868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.357884 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.460781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.460820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.460831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.460847 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.460858 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.564328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.564398 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.564421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.564455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.564483 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.667695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.667776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.667798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.667829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.667851 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.770945 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.771032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.771050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.771076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.771095 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.874144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.874181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.874190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.874202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.874213 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.977403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.977471 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.977489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.977512 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:02 crc kubenswrapper[4810]: I1008 06:33:02.977528 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:02Z","lastTransitionTime":"2025-10-08T06:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.080736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.080820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.080845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.080876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.080893 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.184439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.184488 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.184504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.184538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.184554 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.287489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.287542 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.287553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.287570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.287586 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.390515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.390571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.390582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.390605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.390616 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.494150 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.494200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.494216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.494238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.494253 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.597913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.598046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.598072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.598101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.598122 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.701508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.701581 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.701603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.701634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.701656 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.804531 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.804601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.804618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.804642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.804661 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.911227 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.911292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.911325 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.911351 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:03 crc kubenswrapper[4810]: I1008 06:33:03.911369 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:03Z","lastTransitionTime":"2025-10-08T06:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.014186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.014265 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.014291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.014386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.014410 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.077211 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.077212 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.077301 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:04 crc kubenswrapper[4810]: E1008 06:33:04.077483 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.077561 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:04 crc kubenswrapper[4810]: E1008 06:33:04.077635 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:04 crc kubenswrapper[4810]: E1008 06:33:04.077773 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:04 crc kubenswrapper[4810]: E1008 06:33:04.077935 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.117224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.117253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.117262 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.117291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.117303 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.219953 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.220065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.220089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.220122 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.220147 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.323279 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.323353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.323388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.323422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.323445 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.425825 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.425906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.425923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.425947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.425993 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.529241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.529321 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.529344 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.529376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.529401 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.632794 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.632869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.632882 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.632927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.632935 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.735631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.735705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.735724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.735751 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.735771 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.840232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.840307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.840330 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.840632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.840767 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.943859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.943922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.943939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.943987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.944005 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.980384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.980445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.980454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.980468 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 06:33:04 crc kubenswrapper[4810]: I1008 06:33:04.980477 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T06:33:04Z","lastTransitionTime":"2025-10-08T06:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.049228 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg"] Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.049591 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.052676 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.053604 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.053699 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.056447 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.095597 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fjrrk" podStartSLOduration=64.095568586 podStartE2EDuration="1m4.095568586s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.079133397 +0000 UTC m=+87.713573207" watchObservedRunningTime="2025-10-08 06:33:05.095568586 +0000 UTC m=+87.730008336" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.110087 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podStartSLOduration=64.110058045 podStartE2EDuration="1m4.110058045s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.096113592 +0000 UTC m=+87.730553342" watchObservedRunningTime="2025-10-08 06:33:05.110058045 +0000 UTC m=+87.744497825" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.110326 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gcps7" podStartSLOduration=64.110316592 podStartE2EDuration="1m4.110316592s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.110198649 +0000 UTC m=+87.744638389" watchObservedRunningTime="2025-10-08 06:33:05.110316592 +0000 UTC m=+87.744756372" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.125618 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ktkpc" podStartSLOduration=63.125598271 podStartE2EDuration="1m3.125598271s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.125158139 +0000 UTC m=+87.759597919" watchObservedRunningTime="2025-10-08 06:33:05.125598271 +0000 UTC m=+87.760038021" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.139351 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8ae47a5d-b07d-4425-9bf5-d403aab009d2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.139402 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8ae47a5d-b07d-4425-9bf5-d403aab009d2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.139417 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8ae47a5d-b07d-4425-9bf5-d403aab009d2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.139435 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ae47a5d-b07d-4425-9bf5-d403aab009d2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.139452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ae47a5d-b07d-4425-9bf5-d403aab009d2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.165727 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=8.165701174 podStartE2EDuration="8.165701174s" podCreationTimestamp="2025-10-08 06:32:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.165144299 +0000 UTC m=+87.799584079" watchObservedRunningTime="2025-10-08 06:33:05.165701174 +0000 UTC m=+87.800140944" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.195785 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-gxpv2" podStartSLOduration=64.195757749 podStartE2EDuration="1m4.195757749s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.181379104 +0000 UTC m=+87.815818874" watchObservedRunningTime="2025-10-08 06:33:05.195757749 +0000 UTC m=+87.830197509" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.241021 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8ae47a5d-b07d-4425-9bf5-d403aab009d2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.241080 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8ae47a5d-b07d-4425-9bf5-d403aab009d2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.241120 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8ae47a5d-b07d-4425-9bf5-d403aab009d2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.241145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ae47a5d-b07d-4425-9bf5-d403aab009d2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.241166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ae47a5d-b07d-4425-9bf5-d403aab009d2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.241249 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8ae47a5d-b07d-4425-9bf5-d403aab009d2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.241348 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8ae47a5d-b07d-4425-9bf5-d403aab009d2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.244298 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8ae47a5d-b07d-4425-9bf5-d403aab009d2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.250732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ae47a5d-b07d-4425-9bf5-d403aab009d2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.260178 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ae47a5d-b07d-4425-9bf5-d403aab009d2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w8gmg\" (UID: \"8ae47a5d-b07d-4425-9bf5-d403aab009d2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.266336 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=64.266314718 podStartE2EDuration="1m4.266314718s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.265566328 +0000 UTC m=+87.900006098" watchObservedRunningTime="2025-10-08 06:33:05.266314718 +0000 UTC m=+87.900754488" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.326290 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.326269223 podStartE2EDuration="40.326269223s" podCreationTimestamp="2025-10-08 06:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.325386889 +0000 UTC m=+87.959826669" watchObservedRunningTime="2025-10-08 06:33:05.326269223 +0000 UTC m=+87.960708963" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.369606 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.396357 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-txvr6" podStartSLOduration=64.396337789 podStartE2EDuration="1m4.396337789s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.396151034 +0000 UTC m=+88.030590774" watchObservedRunningTime="2025-10-08 06:33:05.396337789 +0000 UTC m=+88.030777529" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.414950 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=68.414933896 podStartE2EDuration="1m8.414933896s" podCreationTimestamp="2025-10-08 06:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.414755762 +0000 UTC m=+88.049195502" watchObservedRunningTime="2025-10-08 06:33:05.414933896 +0000 UTC m=+88.049373636" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.680632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" event={"ID":"8ae47a5d-b07d-4425-9bf5-d403aab009d2","Type":"ContainerStarted","Data":"68d5da274bc1c5538a5eca6f098958d779f38c488e59268bec79db4da50ef36f"} Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.680678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" event={"ID":"8ae47a5d-b07d-4425-9bf5-d403aab009d2","Type":"ContainerStarted","Data":"e3e69145ea22d84a774c2f5347b21b81d52481879136fc5d76eeb8a50002b61f"} Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.697920 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.697903862 podStartE2EDuration="1m8.697903862s" podCreationTimestamp="2025-10-08 06:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.431302185 +0000 UTC m=+88.065741935" watchObservedRunningTime="2025-10-08 06:33:05.697903862 +0000 UTC m=+88.332343592" Oct 08 06:33:05 crc kubenswrapper[4810]: I1008 06:33:05.698450 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8gmg" podStartSLOduration=64.698445587 podStartE2EDuration="1m4.698445587s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:05.698280493 +0000 UTC m=+88.332720233" watchObservedRunningTime="2025-10-08 06:33:05.698445587 +0000 UTC m=+88.332885327" Oct 08 06:33:06 crc kubenswrapper[4810]: I1008 06:33:06.073259 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:06 crc kubenswrapper[4810]: I1008 06:33:06.073262 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:06 crc kubenswrapper[4810]: I1008 06:33:06.073282 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:06 crc kubenswrapper[4810]: E1008 06:33:06.073978 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:06 crc kubenswrapper[4810]: E1008 06:33:06.074103 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:06 crc kubenswrapper[4810]: I1008 06:33:06.073379 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:06 crc kubenswrapper[4810]: E1008 06:33:06.074248 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:06 crc kubenswrapper[4810]: E1008 06:33:06.074405 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:08 crc kubenswrapper[4810]: I1008 06:33:08.072497 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:08 crc kubenswrapper[4810]: I1008 06:33:08.072567 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:08 crc kubenswrapper[4810]: I1008 06:33:08.072616 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:08 crc kubenswrapper[4810]: I1008 06:33:08.072511 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:08 crc kubenswrapper[4810]: E1008 06:33:08.074513 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:08 crc kubenswrapper[4810]: E1008 06:33:08.074659 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:08 crc kubenswrapper[4810]: E1008 06:33:08.074794 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:08 crc kubenswrapper[4810]: E1008 06:33:08.074983 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:10 crc kubenswrapper[4810]: I1008 06:33:10.072849 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:10 crc kubenswrapper[4810]: E1008 06:33:10.073355 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:10 crc kubenswrapper[4810]: I1008 06:33:10.072929 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:10 crc kubenswrapper[4810]: E1008 06:33:10.073445 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:10 crc kubenswrapper[4810]: I1008 06:33:10.072905 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:10 crc kubenswrapper[4810]: I1008 06:33:10.073057 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:10 crc kubenswrapper[4810]: E1008 06:33:10.073508 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:10 crc kubenswrapper[4810]: E1008 06:33:10.073696 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:12 crc kubenswrapper[4810]: I1008 06:33:12.072931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:12 crc kubenswrapper[4810]: E1008 06:33:12.073077 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:12 crc kubenswrapper[4810]: I1008 06:33:12.073242 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:12 crc kubenswrapper[4810]: I1008 06:33:12.073256 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:12 crc kubenswrapper[4810]: E1008 06:33:12.073399 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:12 crc kubenswrapper[4810]: I1008 06:33:12.073490 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:12 crc kubenswrapper[4810]: E1008 06:33:12.073584 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:12 crc kubenswrapper[4810]: E1008 06:33:12.073687 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:13 crc kubenswrapper[4810]: I1008 06:33:13.073633 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:33:13 crc kubenswrapper[4810]: E1008 06:33:13.073783 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:33:14 crc kubenswrapper[4810]: I1008 06:33:14.072529 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:14 crc kubenswrapper[4810]: I1008 06:33:14.072569 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:14 crc kubenswrapper[4810]: I1008 06:33:14.072644 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:14 crc kubenswrapper[4810]: I1008 06:33:14.072710 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:14 crc kubenswrapper[4810]: E1008 06:33:14.072707 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:14 crc kubenswrapper[4810]: E1008 06:33:14.072832 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:14 crc kubenswrapper[4810]: E1008 06:33:14.072939 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:14 crc kubenswrapper[4810]: E1008 06:33:14.073038 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:16 crc kubenswrapper[4810]: I1008 06:33:16.073439 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:16 crc kubenswrapper[4810]: I1008 06:33:16.073518 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:16 crc kubenswrapper[4810]: I1008 06:33:16.073452 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:16 crc kubenswrapper[4810]: I1008 06:33:16.073653 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:16 crc kubenswrapper[4810]: E1008 06:33:16.073739 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:16 crc kubenswrapper[4810]: E1008 06:33:16.073867 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:16 crc kubenswrapper[4810]: E1008 06:33:16.073960 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:16 crc kubenswrapper[4810]: E1008 06:33:16.074067 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:18 crc kubenswrapper[4810]: I1008 06:33:18.072857 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:18 crc kubenswrapper[4810]: I1008 06:33:18.072886 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:18 crc kubenswrapper[4810]: I1008 06:33:18.072959 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:18 crc kubenswrapper[4810]: I1008 06:33:18.073041 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:18 crc kubenswrapper[4810]: E1008 06:33:18.073916 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:18 crc kubenswrapper[4810]: E1008 06:33:18.074050 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:18 crc kubenswrapper[4810]: E1008 06:33:18.074143 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:18 crc kubenswrapper[4810]: E1008 06:33:18.074361 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:20 crc kubenswrapper[4810]: I1008 06:33:20.073055 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:20 crc kubenswrapper[4810]: I1008 06:33:20.073381 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:20 crc kubenswrapper[4810]: I1008 06:33:20.073437 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:20 crc kubenswrapper[4810]: I1008 06:33:20.073630 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:20 crc kubenswrapper[4810]: E1008 06:33:20.073610 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:20 crc kubenswrapper[4810]: E1008 06:33:20.073766 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:20 crc kubenswrapper[4810]: E1008 06:33:20.073913 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:20 crc kubenswrapper[4810]: E1008 06:33:20.074064 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:20 crc kubenswrapper[4810]: I1008 06:33:20.521457 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:20 crc kubenswrapper[4810]: E1008 06:33:20.521712 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:33:20 crc kubenswrapper[4810]: E1008 06:33:20.521834 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs podName:669c3fc4-7586-4517-83c9-6316b2671329 nodeName:}" failed. No retries permitted until 2025-10-08 06:34:24.521807419 +0000 UTC m=+167.156247169 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs") pod "network-metrics-daemon-9s2m8" (UID: "669c3fc4-7586-4517-83c9-6316b2671329") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 06:33:22 crc kubenswrapper[4810]: I1008 06:33:22.072727 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:22 crc kubenswrapper[4810]: I1008 06:33:22.072610 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:22 crc kubenswrapper[4810]: E1008 06:33:22.072870 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:22 crc kubenswrapper[4810]: I1008 06:33:22.073079 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:22 crc kubenswrapper[4810]: E1008 06:33:22.073164 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:22 crc kubenswrapper[4810]: E1008 06:33:22.073388 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:22 crc kubenswrapper[4810]: I1008 06:33:22.074065 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:22 crc kubenswrapper[4810]: E1008 06:33:22.074427 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:24 crc kubenswrapper[4810]: I1008 06:33:24.073224 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:24 crc kubenswrapper[4810]: I1008 06:33:24.073311 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:24 crc kubenswrapper[4810]: I1008 06:33:24.073342 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:24 crc kubenswrapper[4810]: I1008 06:33:24.073438 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:24 crc kubenswrapper[4810]: E1008 06:33:24.074594 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:24 crc kubenswrapper[4810]: E1008 06:33:24.074764 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:24 crc kubenswrapper[4810]: E1008 06:33:24.074871 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:24 crc kubenswrapper[4810]: E1008 06:33:24.074908 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:26 crc kubenswrapper[4810]: I1008 06:33:26.072528 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:26 crc kubenswrapper[4810]: I1008 06:33:26.072612 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:26 crc kubenswrapper[4810]: E1008 06:33:26.072691 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:26 crc kubenswrapper[4810]: I1008 06:33:26.072529 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:26 crc kubenswrapper[4810]: E1008 06:33:26.072801 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:26 crc kubenswrapper[4810]: I1008 06:33:26.072823 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:26 crc kubenswrapper[4810]: E1008 06:33:26.072934 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:26 crc kubenswrapper[4810]: E1008 06:33:26.073095 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:27 crc kubenswrapper[4810]: I1008 06:33:27.073799 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:33:27 crc kubenswrapper[4810]: E1008 06:33:27.073952 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qkrh9_openshift-ovn-kubernetes(ee7b17fc-599e-4723-80fb-eb248dfe539d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" Oct 08 06:33:28 crc kubenswrapper[4810]: I1008 06:33:28.072435 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:28 crc kubenswrapper[4810]: I1008 06:33:28.072502 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:28 crc kubenswrapper[4810]: I1008 06:33:28.072585 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:28 crc kubenswrapper[4810]: E1008 06:33:28.075500 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:28 crc kubenswrapper[4810]: I1008 06:33:28.075760 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:28 crc kubenswrapper[4810]: E1008 06:33:28.075833 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:28 crc kubenswrapper[4810]: E1008 06:33:28.076028 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:28 crc kubenswrapper[4810]: E1008 06:33:28.076814 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:30 crc kubenswrapper[4810]: I1008 06:33:30.072420 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:30 crc kubenswrapper[4810]: I1008 06:33:30.072537 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:30 crc kubenswrapper[4810]: I1008 06:33:30.072473 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:30 crc kubenswrapper[4810]: I1008 06:33:30.072463 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:30 crc kubenswrapper[4810]: E1008 06:33:30.073338 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:30 crc kubenswrapper[4810]: E1008 06:33:30.073410 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:30 crc kubenswrapper[4810]: E1008 06:33:30.073417 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:30 crc kubenswrapper[4810]: E1008 06:33:30.073554 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:32 crc kubenswrapper[4810]: I1008 06:33:32.073146 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:32 crc kubenswrapper[4810]: I1008 06:33:32.073195 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:32 crc kubenswrapper[4810]: I1008 06:33:32.073196 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:32 crc kubenswrapper[4810]: E1008 06:33:32.073328 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:32 crc kubenswrapper[4810]: I1008 06:33:32.073443 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:32 crc kubenswrapper[4810]: E1008 06:33:32.073474 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:32 crc kubenswrapper[4810]: E1008 06:33:32.073694 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:32 crc kubenswrapper[4810]: E1008 06:33:32.073926 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:34 crc kubenswrapper[4810]: I1008 06:33:34.072380 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:34 crc kubenswrapper[4810]: E1008 06:33:34.072516 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:34 crc kubenswrapper[4810]: I1008 06:33:34.072391 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:34 crc kubenswrapper[4810]: I1008 06:33:34.072584 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:34 crc kubenswrapper[4810]: E1008 06:33:34.072600 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:34 crc kubenswrapper[4810]: I1008 06:33:34.072392 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:34 crc kubenswrapper[4810]: E1008 06:33:34.072854 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:34 crc kubenswrapper[4810]: E1008 06:33:34.072774 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:35 crc kubenswrapper[4810]: I1008 06:33:35.803874 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/1.log" Oct 08 06:33:35 crc kubenswrapper[4810]: I1008 06:33:35.804874 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/0.log" Oct 08 06:33:35 crc kubenswrapper[4810]: I1008 06:33:35.804927 4810 generic.go:334] "Generic (PLEG): container finished" podID="9f654c73-c1a2-4292-abe0-830b46ed68f5" containerID="2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1" exitCode=1 Oct 08 06:33:35 crc kubenswrapper[4810]: I1008 06:33:35.804963 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerDied","Data":"2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1"} Oct 08 06:33:35 crc kubenswrapper[4810]: I1008 06:33:35.805040 4810 scope.go:117] "RemoveContainer" containerID="29b5b6b588ea2bda937ad70a4f2af761d30df1985810095f0126986d43622414" Oct 08 06:33:35 crc kubenswrapper[4810]: I1008 06:33:35.805389 4810 scope.go:117] "RemoveContainer" containerID="2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1" Oct 08 06:33:35 crc kubenswrapper[4810]: E1008 06:33:35.806224 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-txvr6_openshift-multus(9f654c73-c1a2-4292-abe0-830b46ed68f5)\"" pod="openshift-multus/multus-txvr6" podUID="9f654c73-c1a2-4292-abe0-830b46ed68f5" Oct 08 06:33:36 crc kubenswrapper[4810]: I1008 06:33:36.073032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:36 crc kubenswrapper[4810]: I1008 06:33:36.073032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:36 crc kubenswrapper[4810]: I1008 06:33:36.073055 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:36 crc kubenswrapper[4810]: E1008 06:33:36.073577 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:36 crc kubenswrapper[4810]: E1008 06:33:36.073400 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:36 crc kubenswrapper[4810]: I1008 06:33:36.073118 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:36 crc kubenswrapper[4810]: E1008 06:33:36.073740 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:36 crc kubenswrapper[4810]: E1008 06:33:36.073743 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:36 crc kubenswrapper[4810]: I1008 06:33:36.810177 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/1.log" Oct 08 06:33:38 crc kubenswrapper[4810]: I1008 06:33:38.072800 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:38 crc kubenswrapper[4810]: I1008 06:33:38.072837 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:38 crc kubenswrapper[4810]: E1008 06:33:38.075096 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:38 crc kubenswrapper[4810]: I1008 06:33:38.075119 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:38 crc kubenswrapper[4810]: I1008 06:33:38.075230 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:38 crc kubenswrapper[4810]: E1008 06:33:38.075422 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:38 crc kubenswrapper[4810]: E1008 06:33:38.075519 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:38 crc kubenswrapper[4810]: E1008 06:33:38.075669 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:38 crc kubenswrapper[4810]: E1008 06:33:38.100644 4810 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 08 06:33:38 crc kubenswrapper[4810]: E1008 06:33:38.195176 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 06:33:40 crc kubenswrapper[4810]: I1008 06:33:40.072990 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:40 crc kubenswrapper[4810]: I1008 06:33:40.073030 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:40 crc kubenswrapper[4810]: I1008 06:33:40.073100 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:40 crc kubenswrapper[4810]: E1008 06:33:40.073187 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:40 crc kubenswrapper[4810]: I1008 06:33:40.073273 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:40 crc kubenswrapper[4810]: E1008 06:33:40.073444 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:40 crc kubenswrapper[4810]: E1008 06:33:40.073500 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:40 crc kubenswrapper[4810]: E1008 06:33:40.073624 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:41 crc kubenswrapper[4810]: I1008 06:33:41.073274 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:33:41 crc kubenswrapper[4810]: I1008 06:33:41.829937 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/3.log" Oct 08 06:33:41 crc kubenswrapper[4810]: I1008 06:33:41.832926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerStarted","Data":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} Oct 08 06:33:41 crc kubenswrapper[4810]: I1008 06:33:41.833302 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:33:41 crc kubenswrapper[4810]: I1008 06:33:41.872041 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podStartSLOduration=100.8720174 podStartE2EDuration="1m40.8720174s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:33:41.870480279 +0000 UTC m=+124.504920019" watchObservedRunningTime="2025-10-08 06:33:41.8720174 +0000 UTC m=+124.506457150" Oct 08 06:33:42 crc kubenswrapper[4810]: I1008 06:33:42.072263 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:42 crc kubenswrapper[4810]: I1008 06:33:42.072351 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:42 crc kubenswrapper[4810]: E1008 06:33:42.072420 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:42 crc kubenswrapper[4810]: I1008 06:33:42.072445 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:42 crc kubenswrapper[4810]: I1008 06:33:42.072468 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:42 crc kubenswrapper[4810]: E1008 06:33:42.072658 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:42 crc kubenswrapper[4810]: E1008 06:33:42.072767 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:42 crc kubenswrapper[4810]: E1008 06:33:42.072863 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:42 crc kubenswrapper[4810]: I1008 06:33:42.256674 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9s2m8"] Oct 08 06:33:42 crc kubenswrapper[4810]: I1008 06:33:42.835750 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:42 crc kubenswrapper[4810]: E1008 06:33:42.835870 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:43 crc kubenswrapper[4810]: E1008 06:33:43.196444 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 06:33:44 crc kubenswrapper[4810]: I1008 06:33:44.072807 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:44 crc kubenswrapper[4810]: I1008 06:33:44.072929 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:44 crc kubenswrapper[4810]: I1008 06:33:44.073010 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:44 crc kubenswrapper[4810]: E1008 06:33:44.073246 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:44 crc kubenswrapper[4810]: E1008 06:33:44.073386 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:44 crc kubenswrapper[4810]: E1008 06:33:44.073560 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:45 crc kubenswrapper[4810]: I1008 06:33:45.073231 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:45 crc kubenswrapper[4810]: E1008 06:33:45.073356 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:46 crc kubenswrapper[4810]: I1008 06:33:46.073157 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:46 crc kubenswrapper[4810]: I1008 06:33:46.073175 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:46 crc kubenswrapper[4810]: E1008 06:33:46.073393 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:46 crc kubenswrapper[4810]: E1008 06:33:46.073481 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:46 crc kubenswrapper[4810]: I1008 06:33:46.073203 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:46 crc kubenswrapper[4810]: E1008 06:33:46.073601 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:47 crc kubenswrapper[4810]: I1008 06:33:47.072722 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:47 crc kubenswrapper[4810]: E1008 06:33:47.073207 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:48 crc kubenswrapper[4810]: I1008 06:33:48.072561 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:48 crc kubenswrapper[4810]: I1008 06:33:48.073227 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:48 crc kubenswrapper[4810]: E1008 06:33:48.074507 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:48 crc kubenswrapper[4810]: I1008 06:33:48.074543 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:48 crc kubenswrapper[4810]: E1008 06:33:48.074672 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:48 crc kubenswrapper[4810]: E1008 06:33:48.074839 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:48 crc kubenswrapper[4810]: E1008 06:33:48.197160 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 06:33:49 crc kubenswrapper[4810]: I1008 06:33:49.072613 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:49 crc kubenswrapper[4810]: E1008 06:33:49.072807 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:50 crc kubenswrapper[4810]: I1008 06:33:50.072725 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:50 crc kubenswrapper[4810]: I1008 06:33:50.072767 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:50 crc kubenswrapper[4810]: E1008 06:33:50.072899 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:50 crc kubenswrapper[4810]: I1008 06:33:50.073022 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:50 crc kubenswrapper[4810]: E1008 06:33:50.073208 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:50 crc kubenswrapper[4810]: E1008 06:33:50.073358 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:51 crc kubenswrapper[4810]: I1008 06:33:51.073107 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:51 crc kubenswrapper[4810]: E1008 06:33:51.073262 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:51 crc kubenswrapper[4810]: I1008 06:33:51.073336 4810 scope.go:117] "RemoveContainer" containerID="2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1" Oct 08 06:33:51 crc kubenswrapper[4810]: I1008 06:33:51.874548 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/1.log" Oct 08 06:33:51 crc kubenswrapper[4810]: I1008 06:33:51.874850 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerStarted","Data":"97724d3ceaf0c93d52284cdd2d590748da432288355b5517444c117dcb5b06d9"} Oct 08 06:33:52 crc kubenswrapper[4810]: I1008 06:33:52.074184 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:52 crc kubenswrapper[4810]: I1008 06:33:52.074242 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:52 crc kubenswrapper[4810]: E1008 06:33:52.074308 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 06:33:52 crc kubenswrapper[4810]: I1008 06:33:52.074318 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:52 crc kubenswrapper[4810]: E1008 06:33:52.074498 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 06:33:52 crc kubenswrapper[4810]: E1008 06:33:52.074566 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 06:33:53 crc kubenswrapper[4810]: I1008 06:33:53.072634 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:53 crc kubenswrapper[4810]: E1008 06:33:53.072787 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9s2m8" podUID="669c3fc4-7586-4517-83c9-6316b2671329" Oct 08 06:33:54 crc kubenswrapper[4810]: I1008 06:33:54.073170 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:33:54 crc kubenswrapper[4810]: I1008 06:33:54.073259 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:33:54 crc kubenswrapper[4810]: I1008 06:33:54.073366 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:33:54 crc kubenswrapper[4810]: I1008 06:33:54.076005 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 06:33:54 crc kubenswrapper[4810]: I1008 06:33:54.076239 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 06:33:54 crc kubenswrapper[4810]: I1008 06:33:54.076451 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 06:33:54 crc kubenswrapper[4810]: I1008 06:33:54.077156 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 06:33:55 crc kubenswrapper[4810]: I1008 06:33:55.072923 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:33:55 crc kubenswrapper[4810]: I1008 06:33:55.076046 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 08 06:33:55 crc kubenswrapper[4810]: I1008 06:33:55.076666 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.446524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.495846 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jn2mz"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.496756 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.497458 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-56cpr"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.498164 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ch8bp"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.498401 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.498628 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.498914 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nhfv4"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.499301 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.503017 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.503023 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.503634 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.503710 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.503759 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.503907 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.504480 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.505143 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.507056 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sxjjh"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.507146 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.508082 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.511634 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.511669 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.511753 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.511667 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.511824 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.512227 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.512567 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.512830 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.512876 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.513113 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.515092 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.519226 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.520069 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.527083 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.529294 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.530004 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.530777 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.531339 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.532351 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.537184 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.541807 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543178 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543339 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543621 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543747 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543804 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543844 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543920 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.543942 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.544035 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.544153 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.544268 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.544386 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.544502 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.545123 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.545233 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.545331 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.545332 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.545861 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.545982 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.546169 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.585773 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-fq7wx"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.586250 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.586659 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6m85n"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.587055 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6m85n" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.587481 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.587800 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.589921 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kt2pp"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.590633 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.591096 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jn2mz"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.591191 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.591873 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.592478 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.593775 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.594026 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.594317 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.594428 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.594458 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.594631 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.594769 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.594990 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.597766 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.598269 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.598622 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.598801 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.601494 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.602091 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.602734 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604326 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b92a393e-8f4d-4472-9302-98be9693c774-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604362 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/80088141-deec-4bdb-b54a-10fce37e97a0-images\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89rs6\" (UniqueName: \"kubernetes.io/projected/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-kube-api-access-89rs6\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604438 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-dir\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604459 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604507 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b92a393e-8f4d-4472-9302-98be9693c774-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604529 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-node-pullsecrets\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-config\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604605 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwbb\" (UniqueName: \"kubernetes.io/projected/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-kube-api-access-fmwbb\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604627 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-image-import-ca\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604676 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b63477b-8171-4222-815b-16c7b5ba0e13-serving-cert\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604698 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604719 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-oauth-serving-cert\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604688 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604764 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5fd4782-1ec5-49ed-9db0-6a552d621671-audit-dir\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604835 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b92a393e-8f4d-4472-9302-98be9693c774-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604873 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj4ft\" (UniqueName: \"kubernetes.io/projected/b92a393e-8f4d-4472-9302-98be9693c774-kube-api-access-gj4ft\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604915 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.604938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313a8482-046d-4e1c-aebe-ea694d3549b3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605033 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-service-ca-bundle\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w944\" (UniqueName: \"kubernetes.io/projected/80088141-deec-4bdb-b54a-10fce37e97a0-kube-api-access-4w944\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605221 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8457\" (UniqueName: \"kubernetes.io/projected/303e572d-0bdf-4cf2-85bf-976e5ac59130-kube-api-access-g8457\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605276 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-serving-cert\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605297 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-encryption-config\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605341 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-config\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605379 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605385 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-client-ca\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605409 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-trusted-ca\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605430 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-audit-policies\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605441 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605453 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-encryption-config\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605478 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/593a9343-25b7-4fce-b32d-8f683173bee5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lgrl2\" (UID: \"593a9343-25b7-4fce-b32d-8f683173bee5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-etcd-client\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605575 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxqpg\" (UniqueName: \"kubernetes.io/projected/e1f50b1f-b80c-476f-a461-b40ea60f4312-kube-api-access-bxqpg\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605595 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605616 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605652 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-config\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605671 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-client-ca\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605699 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m68mn\" (UniqueName: \"kubernetes.io/projected/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-kube-api-access-m68mn\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605717 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605737 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605776 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88xlc\" (UniqueName: \"kubernetes.io/projected/9b63477b-8171-4222-815b-16c7b5ba0e13-kube-api-access-88xlc\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605792 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605801 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-service-ca\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605822 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4980c5d8-c614-4943-aff6-cb07dac7593f-serving-cert\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-config\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605926 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cfbq\" (UniqueName: \"kubernetes.io/projected/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-kube-api-access-7cfbq\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.605982 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmsmf\" (UniqueName: \"kubernetes.io/projected/c5fd4782-1ec5-49ed-9db0-6a552d621671-kube-api-access-vmsmf\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606005 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80088141-deec-4bdb-b54a-10fce37e97a0-config\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606025 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-audit\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606045 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-policies\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606066 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606086 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-serving-cert\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606111 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-serving-cert\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606131 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-oauth-config\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606129 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606168 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1f50b1f-b80c-476f-a461-b40ea60f4312-serving-cert\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-config\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606210 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606229 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-etcd-client\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606256 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606264 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f779z"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606775 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606276 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606824 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-serving-cert\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606886 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdgwb\" (UniqueName: \"kubernetes.io/projected/26503c29-9886-4ce9-b8f5-b8f03f1fa992-kube-api-access-mdgwb\") pod \"downloads-7954f5f757-6m85n\" (UID: \"26503c29-9886-4ce9-b8f5-b8f03f1fa992\") " pod="openshift-console/downloads-7954f5f757-6m85n" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606917 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-audit-dir\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.606989 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313a8482-046d-4e1c-aebe-ea694d3549b3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607072 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607096 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-config\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzpz6\" (UniqueName: \"kubernetes.io/projected/4980c5d8-c614-4943-aff6-cb07dac7593f-kube-api-access-rzpz6\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607138 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-trusted-ca-bundle\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607153 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-serving-cert\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607189 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/80088141-deec-4bdb-b54a-10fce37e97a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607206 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607224 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607240 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607255 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607274 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkwbp\" (UniqueName: \"kubernetes.io/projected/313a8482-046d-4e1c-aebe-ea694d3549b3-kube-api-access-dkwbp\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607382 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89d8q\" (UniqueName: \"kubernetes.io/projected/593a9343-25b7-4fce-b32d-8f683173bee5-kube-api-access-89d8q\") pod \"cluster-samples-operator-665b6dd947-lgrl2\" (UID: \"593a9343-25b7-4fce-b32d-8f683173bee5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.607949 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.608254 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.609003 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nqgd7"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.609721 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.610168 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.620335 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.648202 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.648665 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.652391 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.652607 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.652852 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.653088 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.653250 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.653710 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.653897 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.653936 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.654194 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.654305 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.654418 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.654526 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.654627 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.654939 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.655163 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.655272 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.656584 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.656693 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.656790 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.656882 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657045 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657119 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657174 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657213 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657187 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657281 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657426 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657529 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657622 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657725 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.657876 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.658007 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.658157 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.658356 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.658464 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.658581 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.658688 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.658853 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.659891 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.660160 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.660248 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.660288 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.686480 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.686989 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nhfv4"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.687060 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-56cpr"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.687639 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.687813 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.688588 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.703125 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.704106 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.704239 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.704787 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.704811 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.705293 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hm95q"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.705643 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.705796 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gpgrc"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.705940 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.706051 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.706095 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.706595 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.706649 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sxjjh"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.708525 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7q29k"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709166 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709277 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709313 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-service-ca-bundle\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709350 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w944\" (UniqueName: \"kubernetes.io/projected/80088141-deec-4bdb-b54a-10fce37e97a0-kube-api-access-4w944\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709397 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8457\" (UniqueName: \"kubernetes.io/projected/303e572d-0bdf-4cf2-85bf-976e5ac59130-kube-api-access-g8457\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709432 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-serving-cert\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709462 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-encryption-config\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709487 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-config\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709512 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-client-ca\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709534 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-trusted-ca\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709564 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-audit-policies\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-encryption-config\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709616 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/593a9343-25b7-4fce-b32d-8f683173bee5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lgrl2\" (UID: \"593a9343-25b7-4fce-b32d-8f683173bee5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709644 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-auth-proxy-config\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709666 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-config\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709689 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbe763ba-a910-4767-a673-7b7a3d76f342-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709714 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709740 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-etcd-client\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709763 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2e808a56-2e57-410d-ace9-cd65a93881d6-profile-collector-cert\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709784 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/402ce4ba-4a2f-46c0-a810-be60006afd09-metrics-tls\") pod \"dns-operator-744455d44c-nqgd7\" (UID: \"402ce4ba-4a2f-46c0-a810-be60006afd09\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709805 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce854742-cd04-4ccc-81cd-895c46b3d566-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709826 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce854742-cd04-4ccc-81cd-895c46b3d566-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709849 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxqpg\" (UniqueName: \"kubernetes.io/projected/e1f50b1f-b80c-476f-a461-b40ea60f4312-kube-api-access-bxqpg\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709916 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709939 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbe763ba-a910-4767-a673-7b7a3d76f342-proxy-tls\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.709989 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-config\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710012 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-client-ca\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710034 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m68mn\" (UniqueName: \"kubernetes.io/projected/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-kube-api-access-m68mn\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710055 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710078 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710099 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r66nr\" (UniqueName: \"kubernetes.io/projected/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-kube-api-access-r66nr\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710121 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88xlc\" (UniqueName: \"kubernetes.io/projected/9b63477b-8171-4222-815b-16c7b5ba0e13-kube-api-access-88xlc\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-service-ca\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710165 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vkm4\" (UniqueName: \"kubernetes.io/projected/472df6a2-664a-447c-9f9b-ad26fd30c300-kube-api-access-7vkm4\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fda92a79-7c7e-4e9d-b67c-5ed1647946db-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4980c5d8-c614-4943-aff6-cb07dac7593f-serving-cert\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710238 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-config\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710261 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cfbq\" (UniqueName: \"kubernetes.io/projected/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-kube-api-access-7cfbq\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710288 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhl4b\" (UniqueName: \"kubernetes.io/projected/c8a4fee4-3426-4a48-8ae5-eec32d6d70ba-kube-api-access-mhl4b\") pod \"migrator-59844c95c7-72rb2\" (UID: \"c8a4fee4-3426-4a48-8ae5-eec32d6d70ba\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710330 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-config\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710351 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fda92a79-7c7e-4e9d-b67c-5ed1647946db-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710375 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmsmf\" (UniqueName: \"kubernetes.io/projected/c5fd4782-1ec5-49ed-9db0-6a552d621671-kube-api-access-vmsmf\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710400 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80088141-deec-4bdb-b54a-10fce37e97a0-config\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-audit\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710443 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-policies\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710488 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-serving-cert\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710512 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd55j\" (UniqueName: \"kubernetes.io/projected/402ce4ba-4a2f-46c0-a810-be60006afd09-kube-api-access-cd55j\") pod \"dns-operator-744455d44c-nqgd7\" (UID: \"402ce4ba-4a2f-46c0-a810-be60006afd09\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710536 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-serving-cert\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710559 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-oauth-config\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2e808a56-2e57-410d-ace9-cd65a93881d6-srv-cert\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710606 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-config\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710615 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1f50b1f-b80c-476f-a461-b40ea60f4312-serving-cert\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6vm6\" (UniqueName: \"kubernetes.io/projected/e2e82017-c4ed-4ac3-bfca-f117264ffaca-kube-api-access-z6vm6\") pod \"multus-admission-controller-857f4d67dd-f779z\" (UID: \"e2e82017-c4ed-4ac3-bfca-f117264ffaca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-config\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710690 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710712 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-etcd-client\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710736 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710758 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710780 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-serving-cert\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710803 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdgwb\" (UniqueName: \"kubernetes.io/projected/26503c29-9886-4ce9-b8f5-b8f03f1fa992-kube-api-access-mdgwb\") pod \"downloads-7954f5f757-6m85n\" (UID: \"26503c29-9886-4ce9-b8f5-b8f03f1fa992\") " pod="openshift-console/downloads-7954f5f757-6m85n" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710825 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-audit-dir\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710847 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313a8482-046d-4e1c-aebe-ea694d3549b3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-config\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710915 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzpz6\" (UniqueName: \"kubernetes.io/projected/4980c5d8-c614-4943-aff6-cb07dac7593f-kube-api-access-rzpz6\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710936 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-trusted-ca-bundle\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710936 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-service-ca-bundle\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.710979 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-serving-cert\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nvb8\" (UniqueName: \"kubernetes.io/projected/2e808a56-2e57-410d-ace9-cd65a93881d6-kube-api-access-5nvb8\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711050 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc27m\" (UniqueName: \"kubernetes.io/projected/bbe763ba-a910-4767-a673-7b7a3d76f342-kube-api-access-pc27m\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711080 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/80088141-deec-4bdb-b54a-10fce37e97a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711096 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711112 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711169 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711186 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-machine-approver-tls\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711228 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkwbp\" (UniqueName: \"kubernetes.io/projected/313a8482-046d-4e1c-aebe-ea694d3549b3-kube-api-access-dkwbp\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711246 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89d8q\" (UniqueName: \"kubernetes.io/projected/593a9343-25b7-4fce-b32d-8f683173bee5-kube-api-access-89d8q\") pod \"cluster-samples-operator-665b6dd947-lgrl2\" (UID: \"593a9343-25b7-4fce-b32d-8f683173bee5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/472df6a2-664a-447c-9f9b-ad26fd30c300-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711282 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711304 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b92a393e-8f4d-4472-9302-98be9693c774-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711321 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/80088141-deec-4bdb-b54a-10fce37e97a0-images\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711339 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89rs6\" (UniqueName: \"kubernetes.io/projected/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-kube-api-access-89rs6\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-dir\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711385 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711404 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/472df6a2-664a-447c-9f9b-ad26fd30c300-images\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwq9f\" (UniqueName: \"kubernetes.io/projected/ce854742-cd04-4ccc-81cd-895c46b3d566-kube-api-access-nwq9f\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711446 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b92a393e-8f4d-4472-9302-98be9693c774-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711464 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-node-pullsecrets\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711467 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-client-ca\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711813 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.712046 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-config\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.711480 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-config\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.712238 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/80088141-deec-4bdb-b54a-10fce37e97a0-images\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.712277 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.712365 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.712948 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.713019 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.713088 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-node-pullsecrets\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.713646 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.713699 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-config\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.714088 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.714258 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.715120 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kt2pp"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.716015 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/80088141-deec-4bdb-b54a-10fce37e97a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.717755 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.720040 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.720133 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-serving-cert\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.721640 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b92a393e-8f4d-4472-9302-98be9693c774-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.722449 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.723270 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b92a393e-8f4d-4472-9302-98be9693c774-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.724283 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-config\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.724568 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-audit-dir\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.724659 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.724720 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-fwdsd"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.725453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-config\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.725549 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-encryption-config\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726070 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-client-ca\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726121 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2e82017-c4ed-4ac3-bfca-f117264ffaca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f779z\" (UID: \"e2e82017-c4ed-4ac3-bfca-f117264ffaca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726156 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmwbb\" (UniqueName: \"kubernetes.io/projected/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-kube-api-access-fmwbb\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-image-import-ca\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726192 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b63477b-8171-4222-815b-16c7b5ba0e13-serving-cert\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726209 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-oauth-serving-cert\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5fd4782-1ec5-49ed-9db0-6a552d621671-audit-dir\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726262 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/472df6a2-664a-447c-9f9b-ad26fd30c300-proxy-tls\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fda92a79-7c7e-4e9d-b67c-5ed1647946db-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726297 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b92a393e-8f4d-4472-9302-98be9693c774-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj4ft\" (UniqueName: \"kubernetes.io/projected/b92a393e-8f4d-4472-9302-98be9693c774-kube-api-access-gj4ft\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726331 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.726346 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313a8482-046d-4e1c-aebe-ea694d3549b3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.727109 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-oauth-config\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.727479 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/313a8482-046d-4e1c-aebe-ea694d3549b3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.742795 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-dir\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.737729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-audit\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.744521 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-policies\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.744786 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.744795 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-etcd-client\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.745695 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-etcd-client\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.744952 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1f50b1f-b80c-476f-a461-b40ea60f4312-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.744987 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-service-ca\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.745063 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.741018 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-serving-cert\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.745400 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-serving-cert\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.745408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.745505 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.745588 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-audit-policies\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.744828 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313a8482-046d-4e1c-aebe-ea694d3549b3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746174 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746203 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746224 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746571 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-config\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6m85n"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746669 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746813 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1f50b1f-b80c-476f-a461-b40ea60f4312-serving-cert\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.746851 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.747189 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.747303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-trusted-ca\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.747423 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5wdms"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.747698 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80088141-deec-4bdb-b54a-10fce37e97a0-config\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.747865 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.747921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.748238 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.748579 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.748740 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5fd4782-1ec5-49ed-9db0-6a552d621671-audit-dir\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.749199 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.749712 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-oauth-serving-cert\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.749941 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.750330 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f779z"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.750370 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-r68jt"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.750567 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-image-import-ca\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.750838 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.751216 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-encryption-config\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.754183 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.754810 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5fd4782-1ec5-49ed-9db0-6a552d621671-serving-cert\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.754880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.755450 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4980c5d8-c614-4943-aff6-cb07dac7593f-serving-cert\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.756120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.756744 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b63477b-8171-4222-815b-16c7b5ba0e13-serving-cert\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.756805 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5fd4782-1ec5-49ed-9db0-6a552d621671-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.756891 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/593a9343-25b7-4fce-b32d-8f683173bee5-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lgrl2\" (UID: \"593a9343-25b7-4fce-b32d-8f683173bee5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.757293 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-serving-cert\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.758127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-trusted-ca-bundle\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.759953 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.760945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.768667 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.775419 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.778648 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.778866 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.779172 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.779224 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.780005 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.780889 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.782487 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.783172 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.784140 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.785353 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ch8bp"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.786828 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fq7wx"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.787864 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nqgd7"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.789054 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.789911 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pmwfq"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.790520 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.791834 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.793302 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.794599 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t8bxf"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.795731 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.795862 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.797417 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hm95q"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.798240 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.798766 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-752k2"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.799554 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-752k2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.800084 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.801372 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-h7bw5"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.802369 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.802531 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.803683 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.804866 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pmwfq"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.806133 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.807220 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-r68jt"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.808843 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7q29k"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.809531 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.810630 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.811677 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.812754 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-752k2"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.813906 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.815051 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.816140 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5wdms"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.817150 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t8bxf"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.818217 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.818429 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.819420 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.820257 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gpgrc"] Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828544 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/472df6a2-664a-447c-9f9b-ad26fd30c300-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828581 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828603 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/472df6a2-664a-447c-9f9b-ad26fd30c300-images\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828639 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwq9f\" (UniqueName: \"kubernetes.io/projected/ce854742-cd04-4ccc-81cd-895c46b3d566-kube-api-access-nwq9f\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828659 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2e82017-c4ed-4ac3-bfca-f117264ffaca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f779z\" (UID: \"e2e82017-c4ed-4ac3-bfca-f117264ffaca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828693 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/472df6a2-664a-447c-9f9b-ad26fd30c300-proxy-tls\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828710 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fda92a79-7c7e-4e9d-b67c-5ed1647946db-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-config\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828788 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbe763ba-a910-4767-a673-7b7a3d76f342-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828811 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q75tf\" (UID: \"12e1e52f-d7f4-4568-a0ac-79ce8102ff96\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-auth-proxy-config\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828850 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce854742-cd04-4ccc-81cd-895c46b3d566-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828869 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce854742-cd04-4ccc-81cd-895c46b3d566-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828891 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2e808a56-2e57-410d-ace9-cd65a93881d6-profile-collector-cert\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828908 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/402ce4ba-4a2f-46c0-a810-be60006afd09-metrics-tls\") pod \"dns-operator-744455d44c-nqgd7\" (UID: \"402ce4ba-4a2f-46c0-a810-be60006afd09\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828932 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.828954 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbe763ba-a910-4767-a673-7b7a3d76f342-proxy-tls\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2r9l\" (UniqueName: \"kubernetes.io/projected/201eaacc-b3ab-4b74-9105-89d0ef91a74e-kube-api-access-b2r9l\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829035 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r66nr\" (UniqueName: \"kubernetes.io/projected/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-kube-api-access-r66nr\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829055 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vkm4\" (UniqueName: \"kubernetes.io/projected/472df6a2-664a-447c-9f9b-ad26fd30c300-kube-api-access-7vkm4\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829076 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fda92a79-7c7e-4e9d-b67c-5ed1647946db-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829117 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhl4b\" (UniqueName: \"kubernetes.io/projected/c8a4fee4-3426-4a48-8ae5-eec32d6d70ba-kube-api-access-mhl4b\") pod \"migrator-59844c95c7-72rb2\" (UID: \"c8a4fee4-3426-4a48-8ae5-eec32d6d70ba\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829139 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-config\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fda92a79-7c7e-4e9d-b67c-5ed1647946db-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829202 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/201eaacc-b3ab-4b74-9105-89d0ef91a74e-serving-cert\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829223 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/201eaacc-b3ab-4b74-9105-89d0ef91a74e-config\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829247 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd55j\" (UniqueName: \"kubernetes.io/projected/402ce4ba-4a2f-46c0-a810-be60006afd09-kube-api-access-cd55j\") pod \"dns-operator-744455d44c-nqgd7\" (UID: \"402ce4ba-4a2f-46c0-a810-be60006afd09\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829268 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2e808a56-2e57-410d-ace9-cd65a93881d6-srv-cert\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6vm6\" (UniqueName: \"kubernetes.io/projected/e2e82017-c4ed-4ac3-bfca-f117264ffaca-kube-api-access-z6vm6\") pod \"multus-admission-controller-857f4d67dd-f779z\" (UID: \"e2e82017-c4ed-4ac3-bfca-f117264ffaca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829323 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss2ql\" (UniqueName: \"kubernetes.io/projected/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-kube-api-access-ss2ql\") pod \"control-plane-machine-set-operator-78cbb6b69f-q75tf\" (UID: \"12e1e52f-d7f4-4568-a0ac-79ce8102ff96\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829365 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nvb8\" (UniqueName: \"kubernetes.io/projected/2e808a56-2e57-410d-ace9-cd65a93881d6-kube-api-access-5nvb8\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc27m\" (UniqueName: \"kubernetes.io/projected/bbe763ba-a910-4767-a673-7b7a3d76f342-kube-api-access-pc27m\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829414 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-machine-approver-tls\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.829682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/472df6a2-664a-447c-9f9b-ad26fd30c300-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.830146 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/472df6a2-664a-447c-9f9b-ad26fd30c300-images\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.830750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fda92a79-7c7e-4e9d-b67c-5ed1647946db-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.830803 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-auth-proxy-config\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.830876 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-config\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.831104 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce854742-cd04-4ccc-81cd-895c46b3d566-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.831636 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbe763ba-a910-4767-a673-7b7a3d76f342-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.832587 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/472df6a2-664a-447c-9f9b-ad26fd30c300-proxy-tls\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.832587 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-machine-approver-tls\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.833842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fda92a79-7c7e-4e9d-b67c-5ed1647946db-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.834086 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce854742-cd04-4ccc-81cd-895c46b3d566-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.834127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2e82017-c4ed-4ac3-bfca-f117264ffaca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f779z\" (UID: \"e2e82017-c4ed-4ac3-bfca-f117264ffaca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.839709 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.859128 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.863915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2e808a56-2e57-410d-ace9-cd65a93881d6-srv-cert\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.879096 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.882126 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2e808a56-2e57-410d-ace9-cd65a93881d6-profile-collector-cert\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.898679 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.918798 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.930361 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q75tf\" (UID: \"12e1e52f-d7f4-4568-a0ac-79ce8102ff96\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.930453 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2r9l\" (UniqueName: \"kubernetes.io/projected/201eaacc-b3ab-4b74-9105-89d0ef91a74e-kube-api-access-b2r9l\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.930566 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/201eaacc-b3ab-4b74-9105-89d0ef91a74e-serving-cert\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.930591 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/201eaacc-b3ab-4b74-9105-89d0ef91a74e-config\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.930625 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss2ql\" (UniqueName: \"kubernetes.io/projected/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-kube-api-access-ss2ql\") pod \"control-plane-machine-set-operator-78cbb6b69f-q75tf\" (UID: \"12e1e52f-d7f4-4568-a0ac-79ce8102ff96\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.938679 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.958791 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.979673 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.982949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/402ce4ba-4a2f-46c0-a810-be60006afd09-metrics-tls\") pod \"dns-operator-744455d44c-nqgd7\" (UID: \"402ce4ba-4a2f-46c0-a810-be60006afd09\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:56 crc kubenswrapper[4810]: I1008 06:33:56.999085 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.018531 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.020838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-config\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.039356 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.042058 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.059213 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.098164 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.119259 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.139649 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.159259 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.180839 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.199533 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.219885 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.239391 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.259423 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.279354 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.301272 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.319295 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.338869 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.343799 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbe763ba-a910-4767-a673-7b7a3d76f342-proxy-tls\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.360150 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.379428 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.399667 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.419780 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.440825 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.487901 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8457\" (UniqueName: \"kubernetes.io/projected/303e572d-0bdf-4cf2-85bf-976e5ac59130-kube-api-access-g8457\") pod \"console-f9d7485db-fq7wx\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.499529 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.505409 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w944\" (UniqueName: \"kubernetes.io/projected/80088141-deec-4bdb-b54a-10fce37e97a0-kube-api-access-4w944\") pod \"machine-api-operator-5694c8668f-jn2mz\" (UID: \"80088141-deec-4bdb-b54a-10fce37e97a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.547677 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m68mn\" (UniqueName: \"kubernetes.io/projected/6b32f41b-1262-49c0-b8f0-4305d5e77dbc-kube-api-access-m68mn\") pod \"openshift-config-operator-7777fb866f-6x4v5\" (UID: \"6b32f41b-1262-49c0-b8f0-4305d5e77dbc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.547907 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.584611 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkwbp\" (UniqueName: \"kubernetes.io/projected/313a8482-046d-4e1c-aebe-ea694d3549b3-kube-api-access-dkwbp\") pod \"openshift-apiserver-operator-796bbdcf4f-nptxs\" (UID: \"313a8482-046d-4e1c-aebe-ea694d3549b3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.599215 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.607294 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89d8q\" (UniqueName: \"kubernetes.io/projected/593a9343-25b7-4fce-b32d-8f683173bee5-kube-api-access-89d8q\") pod \"cluster-samples-operator-665b6dd947-lgrl2\" (UID: \"593a9343-25b7-4fce-b32d-8f683173bee5\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.611464 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.618834 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.667083 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxqpg\" (UniqueName: \"kubernetes.io/projected/e1f50b1f-b80c-476f-a461-b40ea60f4312-kube-api-access-bxqpg\") pod \"authentication-operator-69f744f599-ch8bp\" (UID: \"e1f50b1f-b80c-476f-a461-b40ea60f4312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.679981 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.690046 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzpz6\" (UniqueName: \"kubernetes.io/projected/4980c5d8-c614-4943-aff6-cb07dac7593f-kube-api-access-rzpz6\") pod \"controller-manager-879f6c89f-56cpr\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.704266 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.710825 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.713871 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89rs6\" (UniqueName: \"kubernetes.io/projected/63a2b7dd-cd9b-44f3-b7bb-b37ae920a699-kube-api-access-89rs6\") pod \"console-operator-58897d9998-nhfv4\" (UID: \"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699\") " pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.730120 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.743061 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88xlc\" (UniqueName: \"kubernetes.io/projected/9b63477b-8171-4222-815b-16c7b5ba0e13-kube-api-access-88xlc\") pod \"route-controller-manager-6576b87f9c-7zw2s\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.758801 4810 request.go:700] Waited for 1.012031039s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.759200 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.765120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cfbq\" (UniqueName: \"kubernetes.io/projected/d83a93fa-815c-4ec8-922e-c459c1fc0e7c-kube-api-access-7cfbq\") pod \"apiserver-76f77b778f-kt2pp\" (UID: \"d83a93fa-815c-4ec8-922e-c459c1fc0e7c\") " pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.780490 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.780990 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmsmf\" (UniqueName: \"kubernetes.io/projected/c5fd4782-1ec5-49ed-9db0-6a552d621671-kube-api-access-vmsmf\") pod \"apiserver-7bbb656c7d-whsfl\" (UID: \"c5fd4782-1ec5-49ed-9db0-6a552d621671\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.783807 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.800860 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.814127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmwbb\" (UniqueName: \"kubernetes.io/projected/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-kube-api-access-fmwbb\") pod \"oauth-openshift-558db77b4-sxjjh\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.814849 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.834321 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdgwb\" (UniqueName: \"kubernetes.io/projected/26503c29-9886-4ce9-b8f5-b8f03f1fa992-kube-api-access-mdgwb\") pod \"downloads-7954f5f757-6m85n\" (UID: \"26503c29-9886-4ce9-b8f5-b8f03f1fa992\") " pod="openshift-console/downloads-7954f5f757-6m85n" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.841227 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.841246 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.854987 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.868547 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.870394 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2"] Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.879173 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.899068 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.913130 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fq7wx"] Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.919029 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 06:33:57 crc kubenswrapper[4810]: E1008 06:33:57.931500 4810 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 08 06:33:57 crc kubenswrapper[4810]: E1008 06:33:57.931831 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/201eaacc-b3ab-4b74-9105-89d0ef91a74e-config podName:201eaacc-b3ab-4b74-9105-89d0ef91a74e nodeName:}" failed. No retries permitted until 2025-10-08 06:33:58.431814453 +0000 UTC m=+141.066254183 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/201eaacc-b3ab-4b74-9105-89d0ef91a74e-config") pod "service-ca-operator-777779d784-r68jt" (UID: "201eaacc-b3ab-4b74-9105-89d0ef91a74e") : failed to sync configmap cache: timed out waiting for the condition Oct 08 06:33:57 crc kubenswrapper[4810]: E1008 06:33:57.932128 4810 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 06:33:57 crc kubenswrapper[4810]: E1008 06:33:57.932157 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-control-plane-machine-set-operator-tls podName:12e1e52f-d7f4-4568-a0ac-79ce8102ff96 nodeName:}" failed. No retries permitted until 2025-10-08 06:33:58.432150084 +0000 UTC m=+141.066589824 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-q75tf" (UID: "12e1e52f-d7f4-4568-a0ac-79ce8102ff96") : failed to sync secret cache: timed out waiting for the condition Oct 08 06:33:57 crc kubenswrapper[4810]: E1008 06:33:57.932173 4810 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 06:33:57 crc kubenswrapper[4810]: E1008 06:33:57.932195 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/201eaacc-b3ab-4b74-9105-89d0ef91a74e-serving-cert podName:201eaacc-b3ab-4b74-9105-89d0ef91a74e nodeName:}" failed. No retries permitted until 2025-10-08 06:33:58.432190155 +0000 UTC m=+141.066629895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/201eaacc-b3ab-4b74-9105-89d0ef91a74e-serving-cert") pod "service-ca-operator-777779d784-r68jt" (UID: "201eaacc-b3ab-4b74-9105-89d0ef91a74e") : failed to sync secret cache: timed out waiting for the condition Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.940852 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.961648 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.982654 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.989503 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.996023 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6m85n" Oct 08 06:33:57 crc kubenswrapper[4810]: I1008 06:33:57.999418 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.019522 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.025732 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.039343 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.061772 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.080558 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.099717 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.120274 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.139077 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.160351 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.201109 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b92a393e-8f4d-4472-9302-98be9693c774-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.210984 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.213194 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj4ft\" (UniqueName: \"kubernetes.io/projected/b92a393e-8f4d-4472-9302-98be9693c774-kube-api-access-gj4ft\") pod \"cluster-image-registry-operator-dc59b4c8b-sqn9f\" (UID: \"b92a393e-8f4d-4472-9302-98be9693c774\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:58 crc kubenswrapper[4810]: W1008 06:33:58.218452 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b32f41b_1262_49c0_b8f0_4305d5e77dbc.slice/crio-cd4bb22f885502cdda5fd2492207b5f3e13496098135d69e78841aaee1346b15 WatchSource:0}: Error finding container cd4bb22f885502cdda5fd2492207b5f3e13496098135d69e78841aaee1346b15: Status 404 returned error can't find the container with id cd4bb22f885502cdda5fd2492207b5f3e13496098135d69e78841aaee1346b15 Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.219617 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.244232 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.264346 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.279171 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.296912 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.302262 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 06:33:58 crc kubenswrapper[4810]: W1008 06:33:58.310266 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b63477b_8171_4222_815b_16c7b5ba0e13.slice/crio-d2d94e59c1d8845e19085dcd4b276596f9aebbf17434cfd8ed8a8d3013124f8a WatchSource:0}: Error finding container d2d94e59c1d8845e19085dcd4b276596f9aebbf17434cfd8ed8a8d3013124f8a: Status 404 returned error can't find the container with id d2d94e59c1d8845e19085dcd4b276596f9aebbf17434cfd8ed8a8d3013124f8a Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.319747 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.335142 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sxjjh"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.338889 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.341734 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jn2mz"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.352509 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ch8bp"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.356451 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-56cpr"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.369369 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.379512 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.400513 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.418762 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.427203 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.437896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.439714 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.444265 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nhfv4"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.450861 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/201eaacc-b3ab-4b74-9105-89d0ef91a74e-serving-cert\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.450909 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/201eaacc-b3ab-4b74-9105-89d0ef91a74e-config\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.451026 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q75tf\" (UID: \"12e1e52f-d7f4-4568-a0ac-79ce8102ff96\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.451735 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/201eaacc-b3ab-4b74-9105-89d0ef91a74e-config\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.455791 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q75tf\" (UID: \"12e1e52f-d7f4-4568-a0ac-79ce8102ff96\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.456100 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/201eaacc-b3ab-4b74-9105-89d0ef91a74e-serving-cert\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.459758 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.479472 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.500023 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.506041 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.519009 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.530149 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6m85n"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.536787 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kt2pp"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.538994 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.564280 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.581272 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.605348 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.626202 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.639348 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.659481 4810 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.679846 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.700378 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.733714 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.739611 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.760691 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.777169 4810 request.go:700] Waited for 1.974645875s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.777767 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f"] Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.781502 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 06:33:58 crc kubenswrapper[4810]: W1008 06:33:58.789758 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb92a393e_8f4d_4472_9302_98be9693c774.slice/crio-30722e2bfad00be64b283c4b7632a926335a4e981f976c7e8176089cbf12d350 WatchSource:0}: Error finding container 30722e2bfad00be64b283c4b7632a926335a4e981f976c7e8176089cbf12d350: Status 404 returned error can't find the container with id 30722e2bfad00be64b283c4b7632a926335a4e981f976c7e8176089cbf12d350 Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.799734 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.819259 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.889935 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vkm4\" (UniqueName: \"kubernetes.io/projected/472df6a2-664a-447c-9f9b-ad26fd30c300-kube-api-access-7vkm4\") pod \"machine-config-operator-74547568cd-6dwqm\" (UID: \"472df6a2-664a-447c-9f9b-ad26fd30c300\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.893034 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ee7a22a-3831-4afa-8093-74d09e3c0e4c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-m89fc\" (UID: \"3ee7a22a-3831-4afa-8093-74d09e3c0e4c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.905450 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" event={"ID":"9b63477b-8171-4222-815b-16c7b5ba0e13","Type":"ContainerStarted","Data":"f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.905498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" event={"ID":"9b63477b-8171-4222-815b-16c7b5ba0e13","Type":"ContainerStarted","Data":"d2d94e59c1d8845e19085dcd4b276596f9aebbf17434cfd8ed8a8d3013124f8a"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.905711 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.907178 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fq7wx" event={"ID":"303e572d-0bdf-4cf2-85bf-976e5ac59130","Type":"ContainerStarted","Data":"568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.907294 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fq7wx" event={"ID":"303e572d-0bdf-4cf2-85bf-976e5ac59130","Type":"ContainerStarted","Data":"77a7dcf6b9070130148865aa937ace541edcdd74ad054460aecc7de24ec4ffaa"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.909906 4810 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-7zw2s container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.909980 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" podUID="9b63477b-8171-4222-815b-16c7b5ba0e13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.911004 4810 generic.go:334] "Generic (PLEG): container finished" podID="c5fd4782-1ec5-49ed-9db0-6a552d621671" containerID="45a0ba1daa4624ec80dbb870c09bf6bd1dabd53c86b25f49da75e413596d40db" exitCode=0 Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.911085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" event={"ID":"c5fd4782-1ec5-49ed-9db0-6a552d621671","Type":"ContainerDied","Data":"45a0ba1daa4624ec80dbb870c09bf6bd1dabd53c86b25f49da75e413596d40db"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.911112 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" event={"ID":"c5fd4782-1ec5-49ed-9db0-6a552d621671","Type":"ContainerStarted","Data":"32dd3e64db841f353534a4abb8cc2613441d74b432f8d27021aecbc1954b216f"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.915337 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" event={"ID":"313a8482-046d-4e1c-aebe-ea694d3549b3","Type":"ContainerStarted","Data":"03b98a144490477c7918367d9d6558c0e7ac800336227c5b93060abde5295bfe"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.915366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" event={"ID":"313a8482-046d-4e1c-aebe-ea694d3549b3","Type":"ContainerStarted","Data":"0d239f96add5f308963b99432314ce5614ccdf6db0704148191b4b1025890934"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.917075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" event={"ID":"80088141-deec-4bdb-b54a-10fce37e97a0","Type":"ContainerStarted","Data":"b09b1117b2b5fb8f98bfeb2aa0d4d1b3b437e802f6794c8fc08a81aa3114c45f"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.917173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" event={"ID":"80088141-deec-4bdb-b54a-10fce37e97a0","Type":"ContainerStarted","Data":"ddc6be4998d86ac0d10ea9c98e45fd7fe2cacb55af4c2074dc5b4daf2472a8d8"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.917239 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" event={"ID":"80088141-deec-4bdb-b54a-10fce37e97a0","Type":"ContainerStarted","Data":"1ac706afe2e090e65a7728070d5425d9dd88867b0d7d857ddbc1382454f6cfb6"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.919482 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwq9f\" (UniqueName: \"kubernetes.io/projected/ce854742-cd04-4ccc-81cd-895c46b3d566-kube-api-access-nwq9f\") pod \"openshift-controller-manager-operator-756b6f6bc6-987vk\" (UID: \"ce854742-cd04-4ccc-81cd-895c46b3d566\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.920012 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" event={"ID":"c578e5e6-b81a-490e-ad3c-9bfe81172b1b","Type":"ContainerStarted","Data":"64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.920066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" event={"ID":"c578e5e6-b81a-490e-ad3c-9bfe81172b1b","Type":"ContainerStarted","Data":"81ca4db33a9b5c7e3ebbceac635530bd9d66c3d2a094e27b3e1bffb8bd5625e1"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.923520 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" event={"ID":"4980c5d8-c614-4943-aff6-cb07dac7593f","Type":"ContainerStarted","Data":"ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.923578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" event={"ID":"4980c5d8-c614-4943-aff6-cb07dac7593f","Type":"ContainerStarted","Data":"cc96bd6050602c87e392a5b2376b63cee164e01ef72abb779421d9cbbfa33f02"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.924320 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.926207 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" event={"ID":"d83a93fa-815c-4ec8-922e-c459c1fc0e7c","Type":"ContainerStarted","Data":"82ffb805529b9b86dcaa169d5ed9e37fcb8f00daa05be4bed7b336c9924cc51f"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.926343 4810 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-56cpr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.926382 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" podUID="4980c5d8-c614-4943-aff6-cb07dac7593f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.927813 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" event={"ID":"b92a393e-8f4d-4472-9302-98be9693c774","Type":"ContainerStarted","Data":"30722e2bfad00be64b283c4b7632a926335a4e981f976c7e8176089cbf12d350"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.931028 4810 generic.go:334] "Generic (PLEG): container finished" podID="6b32f41b-1262-49c0-b8f0-4305d5e77dbc" containerID="c8a15c1512a70816d5b76dc8d8780b45f41cc719116f1ae984a24151a6a0c32e" exitCode=0 Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.931086 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" event={"ID":"6b32f41b-1262-49c0-b8f0-4305d5e77dbc","Type":"ContainerDied","Data":"c8a15c1512a70816d5b76dc8d8780b45f41cc719116f1ae984a24151a6a0c32e"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.931107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" event={"ID":"6b32f41b-1262-49c0-b8f0-4305d5e77dbc","Type":"ContainerStarted","Data":"cd4bb22f885502cdda5fd2492207b5f3e13496098135d69e78841aaee1346b15"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.933699 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.938349 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc27m\" (UniqueName: \"kubernetes.io/projected/bbe763ba-a910-4767-a673-7b7a3d76f342-kube-api-access-pc27m\") pod \"machine-config-controller-84d6567774-rx5cd\" (UID: \"bbe763ba-a910-4767-a673-7b7a3d76f342\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.939161 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" event={"ID":"e1f50b1f-b80c-476f-a461-b40ea60f4312","Type":"ContainerStarted","Data":"ac7a125c249c70b4045d18dbdabc980133174e0a7c36a6e0d0721f3ebf2eec4e"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.939293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" event={"ID":"e1f50b1f-b80c-476f-a461-b40ea60f4312","Type":"ContainerStarted","Data":"4616a64d4078da5ed78a88dc082c34c9a8a7f8bdd587b3e0105da8b15c5b6e72"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.945016 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" event={"ID":"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699","Type":"ContainerStarted","Data":"b305e14a6c3624f1d7db214acce60ae8086108b3930f3daee57bffc1055fce93"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.945067 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" event={"ID":"63a2b7dd-cd9b-44f3-b7bb-b37ae920a699","Type":"ContainerStarted","Data":"0955d89426e7ab71643a5d1a7b83fab0511ce5edb96acf2f1a98956159142234"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.945429 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.947134 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6m85n" event={"ID":"26503c29-9886-4ce9-b8f5-b8f03f1fa992","Type":"ContainerStarted","Data":"a9dfd9bc437c327298c8b36e785b2b74a74f00d5a0f4b84033749ff83f310a1f"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.947170 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6m85n" event={"ID":"26503c29-9886-4ce9-b8f5-b8f03f1fa992","Type":"ContainerStarted","Data":"68a37706f1e3397fb841b5e98ca4ef57a1a97815e53face37fc3ecf1d8732cb2"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.947439 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6m85n" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.949351 4810 patch_prober.go:28] interesting pod/console-operator-58897d9998-nhfv4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.949391 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" podUID="63a2b7dd-cd9b-44f3-b7bb-b37ae920a699" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.949591 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m85n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.949642 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6m85n" podUID="26503c29-9886-4ce9-b8f5-b8f03f1fa992" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.949767 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.950391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" event={"ID":"593a9343-25b7-4fce-b32d-8f683173bee5","Type":"ContainerStarted","Data":"f7ee0ee0b59e7cdd8712d7ac944def0c061052b02cad8f6a5ecff8ce0ab09418"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.950418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" event={"ID":"593a9343-25b7-4fce-b32d-8f683173bee5","Type":"ContainerStarted","Data":"641af7e726df7fab075d38c9ee972126a6676ee722bff1e975ad0b3617629883"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.950430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" event={"ID":"593a9343-25b7-4fce-b32d-8f683173bee5","Type":"ContainerStarted","Data":"3fec3963a856a0fb568304275716db3ad20272129e21f04cc130150e4717f8fe"} Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.957923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhl4b\" (UniqueName: \"kubernetes.io/projected/c8a4fee4-3426-4a48-8ae5-eec32d6d70ba-kube-api-access-mhl4b\") pod \"migrator-59844c95c7-72rb2\" (UID: \"c8a4fee4-3426-4a48-8ae5-eec32d6d70ba\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.983474 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd55j\" (UniqueName: \"kubernetes.io/projected/402ce4ba-4a2f-46c0-a810-be60006afd09-kube-api-access-cd55j\") pod \"dns-operator-744455d44c-nqgd7\" (UID: \"402ce4ba-4a2f-46c0-a810-be60006afd09\") " pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.995809 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" Oct 08 06:33:58 crc kubenswrapper[4810]: I1008 06:33:58.996247 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r66nr\" (UniqueName: \"kubernetes.io/projected/3868f3e7-62b6-4ba7-b4a7-69192183f6ed-kube-api-access-r66nr\") pod \"machine-approver-56656f9798-wcrhl\" (UID: \"3868f3e7-62b6-4ba7-b4a7-69192183f6ed\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.009888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.017586 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.031774 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6vm6\" (UniqueName: \"kubernetes.io/projected/e2e82017-c4ed-4ac3-bfca-f117264ffaca-kube-api-access-z6vm6\") pod \"multus-admission-controller-857f4d67dd-f779z\" (UID: \"e2e82017-c4ed-4ac3-bfca-f117264ffaca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.047413 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nvb8\" (UniqueName: \"kubernetes.io/projected/2e808a56-2e57-410d-ace9-cd65a93881d6-kube-api-access-5nvb8\") pod \"catalog-operator-68c6474976-wfshr\" (UID: \"2e808a56-2e57-410d-ace9-cd65a93881d6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.065644 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fda92a79-7c7e-4e9d-b67c-5ed1647946db-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2jfkb\" (UID: \"fda92a79-7c7e-4e9d-b67c-5ed1647946db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.075518 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2r9l\" (UniqueName: \"kubernetes.io/projected/201eaacc-b3ab-4b74-9105-89d0ef91a74e-kube-api-access-b2r9l\") pod \"service-ca-operator-777779d784-r68jt\" (UID: \"201eaacc-b3ab-4b74-9105-89d0ef91a74e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.094154 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.106040 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss2ql\" (UniqueName: \"kubernetes.io/projected/12e1e52f-d7f4-4568-a0ac-79ce8102ff96-kube-api-access-ss2ql\") pod \"control-plane-machine-set-operator-78cbb6b69f-q75tf\" (UID: \"12e1e52f-d7f4-4568-a0ac-79ce8102ff96\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173076 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfr2\" (UniqueName: \"kubernetes.io/projected/9d427010-0272-41f7-9883-4ef48df11e53-kube-api-access-8dfr2\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173429 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f25e194-840d-4f74-a2e1-b0a81f4bb420-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173535 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fe30b2-8748-4816-a44e-8428c396e209-config\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173572 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-certificates\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173593 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a714d476-65eb-4e18-a0ba-e874bae2ca79-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173609 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/26a37cc7-419b-489d-aa13-ec53651e212b-node-bootstrap-token\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173644 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-bound-sa-token\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw9hb\" (UniqueName: \"kubernetes.io/projected/ba1edb88-a5c3-4e50-8c6f-071554e22500-kube-api-access-zw9hb\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173719 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbql5\" (UniqueName: \"kubernetes.io/projected/fa8545fa-ddc1-4af9-8095-f22465be5ed7-kube-api-access-lbql5\") pod \"package-server-manager-789f6589d5-h2zzg\" (UID: \"fa8545fa-ddc1-4af9-8095-f22465be5ed7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173734 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dxbm\" (UniqueName: \"kubernetes.io/projected/646d90cc-1f56-4fd1-bd7f-938021eeb92a-kube-api-access-5dxbm\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173788 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-serving-cert\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173806 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173822 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-socket-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173847 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkzxq\" (UniqueName: \"kubernetes.io/projected/26a37cc7-419b-489d-aa13-ec53651e212b-kube-api-access-wkzxq\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173862 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkgjh\" (UniqueName: \"kubernetes.io/projected/2f25e194-840d-4f74-a2e1-b0a81f4bb420-kube-api-access-jkgjh\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173897 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp2g6\" (UniqueName: \"kubernetes.io/projected/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-kube-api-access-pp2g6\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173913 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-csi-data-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.173978 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174026 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-config\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174054 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-registration-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174101 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b125d88d-22fa-45cc-bfa0-564a6b9eb619-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174127 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-apiservice-cert\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174143 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-webhook-cert\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174158 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a714d476-65eb-4e18-a0ba-e874bae2ca79-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbz8m\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-kube-api-access-vbz8m\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174198 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-service-ca\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174260 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skxk6\" (UniqueName: \"kubernetes.io/projected/d1494115-1173-4a30-a886-bc2fa4fb97ab-kube-api-access-skxk6\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjhww\" (UniqueName: \"kubernetes.io/projected/74c3e024-0dfe-4f9c-930e-70896c3b8110-kube-api-access-hjhww\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174352 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5651941c-e009-48b5-b824-69351bb54c6d-config-volume\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174369 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa8545fa-ddc1-4af9-8095-f22465be5ed7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h2zzg\" (UID: \"fa8545fa-ddc1-4af9-8095-f22465be5ed7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.174418 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-client\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.175562 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-default-certificate\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.176017 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86fe30b2-8748-4816-a44e-8428c396e209-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.176056 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9xds\" (UniqueName: \"kubernetes.io/projected/53c6fda7-90ca-40b7-ba7d-2b182a090130-kube-api-access-m9xds\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.176133 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d427010-0272-41f7-9883-4ef48df11e53-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.176463 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:33:59.676447302 +0000 UTC m=+142.310887042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.176769 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1494115-1173-4a30-a886-bc2fa4fb97ab-config-volume\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.177337 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f25e194-840d-4f74-a2e1-b0a81f4bb420-srv-cert\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.177437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d427010-0272-41f7-9883-4ef48df11e53-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.177458 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5651941c-e009-48b5-b824-69351bb54c6d-secret-volume\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.177482 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bg2h\" (UniqueName: \"kubernetes.io/projected/5651941c-e009-48b5-b824-69351bb54c6d-kube-api-access-5bg2h\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.178548 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-tmpfs\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.179846 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-trusted-ca\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.179887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-metrics-certs\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.180682 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7977\" (UniqueName: \"kubernetes.io/projected/b125d88d-22fa-45cc-bfa0-564a6b9eb619-kube-api-access-w7977\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.181154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86fe30b2-8748-4816-a44e-8428c396e209-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.181322 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.189947 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-tls\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194231 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/53c6fda7-90ca-40b7-ba7d-2b182a090130-signing-key\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194323 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-mountpoint-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194372 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwd6m\" (UniqueName: \"kubernetes.io/projected/701ecb4f-1521-41ce-8294-3e2ec4a1f1a4-kube-api-access-lwd6m\") pod \"ingress-canary-pmwfq\" (UID: \"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4\") " pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/701ecb4f-1521-41ce-8294-3e2ec4a1f1a4-cert\") pod \"ingress-canary-pmwfq\" (UID: \"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4\") " pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d1494115-1173-4a30-a886-bc2fa4fb97ab-metrics-tls\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194556 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-plugins-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194587 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b125d88d-22fa-45cc-bfa0-564a6b9eb619-trusted-ca\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/26a37cc7-419b-489d-aa13-ec53651e212b-certs\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194688 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z567\" (UniqueName: \"kubernetes.io/projected/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-kube-api-access-7z567\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194733 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-stats-auth\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194802 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b125d88d-22fa-45cc-bfa0-564a6b9eb619-metrics-tls\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194871 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646d90cc-1f56-4fd1-bd7f-938021eeb92a-service-ca-bundle\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194942 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-ca\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.194990 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/53c6fda7-90ca-40b7-ba7d-2b182a090130-signing-cabundle\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.224050 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.258749 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.262973 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" Oct 08 06:33:59 crc kubenswrapper[4810]: W1008 06:33:59.274623 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3868f3e7_62b6_4ba7_b4a7_69192183f6ed.slice/crio-426831c8045d642d26eb5cadefb2957b3146046ccf0962c273f7d9c9306746e5 WatchSource:0}: Error finding container 426831c8045d642d26eb5cadefb2957b3146046ccf0962c273f7d9c9306746e5: Status 404 returned error can't find the container with id 426831c8045d642d26eb5cadefb2957b3146046ccf0962c273f7d9c9306746e5 Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.274910 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.290724 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.296907 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297250 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-certificates\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297290 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a714d476-65eb-4e18-a0ba-e874bae2ca79-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297310 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/26a37cc7-419b-489d-aa13-ec53651e212b-node-bootstrap-token\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297335 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-bound-sa-token\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297359 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw9hb\" (UniqueName: \"kubernetes.io/projected/ba1edb88-a5c3-4e50-8c6f-071554e22500-kube-api-access-zw9hb\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbql5\" (UniqueName: \"kubernetes.io/projected/fa8545fa-ddc1-4af9-8095-f22465be5ed7-kube-api-access-lbql5\") pod \"package-server-manager-789f6589d5-h2zzg\" (UID: \"fa8545fa-ddc1-4af9-8095-f22465be5ed7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297412 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dxbm\" (UniqueName: \"kubernetes.io/projected/646d90cc-1f56-4fd1-bd7f-938021eeb92a-kube-api-access-5dxbm\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297437 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-serving-cert\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297468 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkzxq\" (UniqueName: \"kubernetes.io/projected/26a37cc7-419b-489d-aa13-ec53651e212b-kube-api-access-wkzxq\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297487 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-socket-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkgjh\" (UniqueName: \"kubernetes.io/projected/2f25e194-840d-4f74-a2e1-b0a81f4bb420-kube-api-access-jkgjh\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297538 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp2g6\" (UniqueName: \"kubernetes.io/projected/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-kube-api-access-pp2g6\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297553 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-csi-data-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-config\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297611 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-registration-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b125d88d-22fa-45cc-bfa0-564a6b9eb619-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297662 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-apiservice-cert\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297681 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a714d476-65eb-4e18-a0ba-e874bae2ca79-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297699 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbz8m\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-kube-api-access-vbz8m\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297717 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-webhook-cert\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297734 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-service-ca\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297770 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skxk6\" (UniqueName: \"kubernetes.io/projected/d1494115-1173-4a30-a886-bc2fa4fb97ab-kube-api-access-skxk6\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297786 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjhww\" (UniqueName: \"kubernetes.io/projected/74c3e024-0dfe-4f9c-930e-70896c3b8110-kube-api-access-hjhww\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5651941c-e009-48b5-b824-69351bb54c6d-config-volume\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297824 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa8545fa-ddc1-4af9-8095-f22465be5ed7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h2zzg\" (UID: \"fa8545fa-ddc1-4af9-8095-f22465be5ed7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297843 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-client\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297861 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-default-certificate\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297882 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86fe30b2-8748-4816-a44e-8428c396e209-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9xds\" (UniqueName: \"kubernetes.io/projected/53c6fda7-90ca-40b7-ba7d-2b182a090130-kube-api-access-m9xds\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297924 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d427010-0272-41f7-9883-4ef48df11e53-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297941 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1494115-1173-4a30-a886-bc2fa4fb97ab-config-volume\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297977 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f25e194-840d-4f74-a2e1-b0a81f4bb420-srv-cert\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.297997 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d427010-0272-41f7-9883-4ef48df11e53-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298014 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5651941c-e009-48b5-b824-69351bb54c6d-secret-volume\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298030 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bg2h\" (UniqueName: \"kubernetes.io/projected/5651941c-e009-48b5-b824-69351bb54c6d-kube-api-access-5bg2h\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298046 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-tmpfs\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298067 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-trusted-ca\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-metrics-certs\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7977\" (UniqueName: \"kubernetes.io/projected/b125d88d-22fa-45cc-bfa0-564a6b9eb619-kube-api-access-w7977\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298174 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86fe30b2-8748-4816-a44e-8428c396e209-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298214 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298232 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-tls\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298249 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/53c6fda7-90ca-40b7-ba7d-2b182a090130-signing-key\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298266 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-mountpoint-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298285 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwd6m\" (UniqueName: \"kubernetes.io/projected/701ecb4f-1521-41ce-8294-3e2ec4a1f1a4-kube-api-access-lwd6m\") pod \"ingress-canary-pmwfq\" (UID: \"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4\") " pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/701ecb4f-1521-41ce-8294-3e2ec4a1f1a4-cert\") pod \"ingress-canary-pmwfq\" (UID: \"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4\") " pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298331 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d1494115-1173-4a30-a886-bc2fa4fb97ab-metrics-tls\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-plugins-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298379 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b125d88d-22fa-45cc-bfa0-564a6b9eb619-trusted-ca\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z567\" (UniqueName: \"kubernetes.io/projected/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-kube-api-access-7z567\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298416 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/26a37cc7-419b-489d-aa13-ec53651e212b-certs\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298434 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-stats-auth\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298456 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b125d88d-22fa-45cc-bfa0-564a6b9eb619-metrics-tls\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298475 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-ca\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298492 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646d90cc-1f56-4fd1-bd7f-938021eeb92a-service-ca-bundle\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/53c6fda7-90ca-40b7-ba7d-2b182a090130-signing-cabundle\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298540 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfr2\" (UniqueName: \"kubernetes.io/projected/9d427010-0272-41f7-9883-4ef48df11e53-kube-api-access-8dfr2\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298556 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f25e194-840d-4f74-a2e1-b0a81f4bb420-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.298588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fe30b2-8748-4816-a44e-8428c396e209-config\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.299021 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:33:59.798948202 +0000 UTC m=+142.433387942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.299842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fe30b2-8748-4816-a44e-8428c396e209-config\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.300642 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-certificates\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.302123 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-service-ca\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.303710 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a714d476-65eb-4e18-a0ba-e874bae2ca79-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.305628 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5651941c-e009-48b5-b824-69351bb54c6d-config-volume\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.305701 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1494115-1173-4a30-a886-bc2fa4fb97ab-config-volume\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.305729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-plugins-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.306075 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646d90cc-1f56-4fd1-bd7f-938021eeb92a-service-ca-bundle\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.308438 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/53c6fda7-90ca-40b7-ba7d-2b182a090130-signing-cabundle\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.308941 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d427010-0272-41f7-9883-4ef48df11e53-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.309313 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-ca\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.310373 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-tmpfs\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.311271 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-socket-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.311537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-csi-data-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.311768 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-trusted-ca\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.313674 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b125d88d-22fa-45cc-bfa0-564a6b9eb619-trusted-ca\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.313843 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.314064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d1494115-1173-4a30-a886-bc2fa4fb97ab-metrics-tls\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.314194 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-registration-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.315198 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-config\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.315269 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba1edb88-a5c3-4e50-8c6f-071554e22500-mountpoint-dir\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.316642 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f25e194-840d-4f74-a2e1-b0a81f4bb420-srv-cert\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.316938 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-default-certificate\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.322110 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5651941c-e009-48b5-b824-69351bb54c6d-secret-volume\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.323400 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-serving-cert\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.323431 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-webhook-cert\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.323739 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86fe30b2-8748-4816-a44e-8428c396e209-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.329623 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa8545fa-ddc1-4af9-8095-f22465be5ed7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h2zzg\" (UID: \"fa8545fa-ddc1-4af9-8095-f22465be5ed7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.336593 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/53c6fda7-90ca-40b7-ba7d-2b182a090130-signing-key\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.337111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/701ecb4f-1521-41ce-8294-3e2ec4a1f1a4-cert\") pod \"ingress-canary-pmwfq\" (UID: \"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4\") " pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.337250 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-apiservice-cert\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.337595 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b125d88d-22fa-45cc-bfa0-564a6b9eb619-metrics-tls\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.337828 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-stats-auth\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.339180 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f25e194-840d-4f74-a2e1-b0a81f4bb420-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.341070 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/646d90cc-1f56-4fd1-bd7f-938021eeb92a-metrics-certs\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.343574 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bg2h\" (UniqueName: \"kubernetes.io/projected/5651941c-e009-48b5-b824-69351bb54c6d-kube-api-access-5bg2h\") pod \"collect-profiles-29331750-s57sb\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.343649 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d427010-0272-41f7-9883-4ef48df11e53-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.348699 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-etcd-client\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.349941 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.350295 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a714d476-65eb-4e18-a0ba-e874bae2ca79-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.350888 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-tls\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.353471 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/26a37cc7-419b-489d-aa13-ec53651e212b-node-bootstrap-token\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.355114 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/26a37cc7-419b-489d-aa13-ec53651e212b-certs\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.388922 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.389000 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkzxq\" (UniqueName: \"kubernetes.io/projected/26a37cc7-419b-489d-aa13-ec53651e212b-kube-api-access-wkzxq\") pod \"machine-config-server-h7bw5\" (UID: \"26a37cc7-419b-489d-aa13-ec53651e212b\") " pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.399234 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-bound-sa-token\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.403079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.403981 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:33:59.903919753 +0000 UTC m=+142.538359493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.408623 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.415915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw9hb\" (UniqueName: \"kubernetes.io/projected/ba1edb88-a5c3-4e50-8c6f-071554e22500-kube-api-access-zw9hb\") pod \"csi-hostpathplugin-t8bxf\" (UID: \"ba1edb88-a5c3-4e50-8c6f-071554e22500\") " pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.439948 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skxk6\" (UniqueName: \"kubernetes.io/projected/d1494115-1173-4a30-a886-bc2fa4fb97ab-kube-api-access-skxk6\") pod \"dns-default-752k2\" (UID: \"d1494115-1173-4a30-a886-bc2fa4fb97ab\") " pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.459358 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbql5\" (UniqueName: \"kubernetes.io/projected/fa8545fa-ddc1-4af9-8095-f22465be5ed7-kube-api-access-lbql5\") pod \"package-server-manager-789f6589d5-h2zzg\" (UID: \"fa8545fa-ddc1-4af9-8095-f22465be5ed7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.462764 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.471411 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-752k2" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.481702 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-h7bw5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.485176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjhww\" (UniqueName: \"kubernetes.io/projected/74c3e024-0dfe-4f9c-930e-70896c3b8110-kube-api-access-hjhww\") pod \"marketplace-operator-79b997595-7q29k\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.499270 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dxbm\" (UniqueName: \"kubernetes.io/projected/646d90cc-1f56-4fd1-bd7f-938021eeb92a-kube-api-access-5dxbm\") pod \"router-default-5444994796-fwdsd\" (UID: \"646d90cc-1f56-4fd1-bd7f-938021eeb92a\") " pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.504077 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.505761 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.005738299 +0000 UTC m=+142.640178039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.517493 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfr2\" (UniqueName: \"kubernetes.io/projected/9d427010-0272-41f7-9883-4ef48df11e53-kube-api-access-8dfr2\") pod \"kube-storage-version-migrator-operator-b67b599dd-q6wf9\" (UID: \"9d427010-0272-41f7-9883-4ef48df11e53\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.538093 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbz8m\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-kube-api-access-vbz8m\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.557762 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z567\" (UniqueName: \"kubernetes.io/projected/2ba1679f-10be-4f8f-843b-0ca4bec2aaa5-kube-api-access-7z567\") pod \"packageserver-d55dfcdfc-q9sq8\" (UID: \"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.576595 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkgjh\" (UniqueName: \"kubernetes.io/projected/2f25e194-840d-4f74-a2e1-b0a81f4bb420-kube-api-access-jkgjh\") pod \"olm-operator-6b444d44fb-2xp4l\" (UID: \"2f25e194-840d-4f74-a2e1-b0a81f4bb420\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.586921 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.599566 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.605018 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp2g6\" (UniqueName: \"kubernetes.io/projected/315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8-kube-api-access-pp2g6\") pod \"etcd-operator-b45778765-hm95q\" (UID: \"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.605677 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.608155 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.608461 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.108447495 +0000 UTC m=+142.742887235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.615131 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.630709 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.633680 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7977\" (UniqueName: \"kubernetes.io/projected/b125d88d-22fa-45cc-bfa0-564a6b9eb619-kube-api-access-w7977\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.650244 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.651666 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.652156 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.655481 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwd6m\" (UniqueName: \"kubernetes.io/projected/701ecb4f-1521-41ce-8294-3e2ec4a1f1a4-kube-api-access-lwd6m\") pod \"ingress-canary-pmwfq\" (UID: \"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4\") " pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.667549 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.674572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/86fe30b2-8748-4816-a44e-8428c396e209-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z7t57\" (UID: \"86fe30b2-8748-4816-a44e-8428c396e209\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.697771 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-r68jt"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.707670 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.709613 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.709767 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.209747374 +0000 UTC m=+142.844187104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.712754 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.713257 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.213244469 +0000 UTC m=+142.847684209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.714441 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b125d88d-22fa-45cc-bfa0-564a6b9eb619-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5jc7h\" (UID: \"b125d88d-22fa-45cc-bfa0-564a6b9eb619\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.718121 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.720454 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9xds\" (UniqueName: \"kubernetes.io/projected/53c6fda7-90ca-40b7-ba7d-2b182a090130-kube-api-access-m9xds\") pod \"service-ca-9c57cc56f-5wdms\" (UID: \"53c6fda7-90ca-40b7-ba7d-2b182a090130\") " pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.729405 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.735763 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f779z"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.736019 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pmwfq" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.813424 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.813676 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.313644889 +0000 UTC m=+142.948084629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.813745 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.814196 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.314180607 +0000 UTC m=+142.948620347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: W1008 06:33:59.832671 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod201eaacc_b3ab_4b74_9105_89d0ef91a74e.slice/crio-8fb0803036c57313a805c5f4d1be007cfcaef4b06fe363d402a3cc227b7c956f WatchSource:0}: Error finding container 8fb0803036c57313a805c5f4d1be007cfcaef4b06fe363d402a3cc227b7c956f: Status 404 returned error can't find the container with id 8fb0803036c57313a805c5f4d1be007cfcaef4b06fe363d402a3cc227b7c956f Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.837776 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb"] Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.915005 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:33:59 crc kubenswrapper[4810]: E1008 06:33:59.915362 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.415348721 +0000 UTC m=+143.049788461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.957315 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" event={"ID":"472df6a2-664a-447c-9f9b-ad26fd30c300","Type":"ContainerStarted","Data":"4270090d03cec963accb77d23ee7bb1a6033df9f3e163367212dfa17855108b7"} Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.958866 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" event={"ID":"6b32f41b-1262-49c0-b8f0-4305d5e77dbc","Type":"ContainerStarted","Data":"d68c66de0e30fd7a2349c1ad2871c6ea41b0560efb10bbeff094565b89f398ca"} Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.959492 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.964217 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" event={"ID":"ce854742-cd04-4ccc-81cd-895c46b3d566","Type":"ContainerStarted","Data":"319927334da356256381a74e60d1b46b1688608691366c7d9fac1267ae372aba"} Oct 08 06:33:59 crc kubenswrapper[4810]: W1008 06:33:59.967202 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbe763ba_a910_4767_a673_7b7a3d76f342.slice/crio-82ef569cba4c187e10ea095a0bef532f1ab97378e79498051cde85c0c39bda2c WatchSource:0}: Error finding container 82ef569cba4c187e10ea095a0bef532f1ab97378e79498051cde85c0c39bda2c: Status 404 returned error can't find the container with id 82ef569cba4c187e10ea095a0bef532f1ab97378e79498051cde85c0c39bda2c Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.968374 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-fwdsd" event={"ID":"646d90cc-1f56-4fd1-bd7f-938021eeb92a","Type":"ContainerStarted","Data":"1bf61a70804007563519e06231bad6393c9b98afea8ebb2c74766c11bb780ac5"} Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.970482 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.974589 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.975728 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" event={"ID":"3868f3e7-62b6-4ba7-b4a7-69192183f6ed","Type":"ContainerStarted","Data":"426831c8045d642d26eb5cadefb2957b3146046ccf0962c273f7d9c9306746e5"} Oct 08 06:33:59 crc kubenswrapper[4810]: W1008 06:33:59.976706 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2e82017_c4ed_4ac3_bfca_f117264ffaca.slice/crio-9a74e3da0d43842109bb5b99a3b228d5fa3394491f2fc00d0aa5cbc855a2efc5 WatchSource:0}: Error finding container 9a74e3da0d43842109bb5b99a3b228d5fa3394491f2fc00d0aa5cbc855a2efc5: Status 404 returned error can't find the container with id 9a74e3da0d43842109bb5b99a3b228d5fa3394491f2fc00d0aa5cbc855a2efc5 Oct 08 06:33:59 crc kubenswrapper[4810]: I1008 06:33:59.980499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" event={"ID":"b92a393e-8f4d-4472-9302-98be9693c774","Type":"ContainerStarted","Data":"07a93cb0f7b9eb5b809f7cf3142b940162ccf659e6ae5c0cea92653a12c5523c"} Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:33:59.999500 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" event={"ID":"3ee7a22a-3831-4afa-8093-74d09e3c0e4c","Type":"ContainerStarted","Data":"c2d6a7608d807170381bbcee708430a11c4c3380157416dbd8f3223f03c83c5a"} Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.003633 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.023887 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.024292 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.524279853 +0000 UTC m=+143.158719593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.024853 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" event={"ID":"c5fd4782-1ec5-49ed-9db0-6a552d621671","Type":"ContainerStarted","Data":"07a406cfd31918b04a7765047413f7cd1f96da43551c104fcf0bad7da70c98ab"} Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.026899 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" event={"ID":"201eaacc-b3ab-4b74-9105-89d0ef91a74e","Type":"ContainerStarted","Data":"8fb0803036c57313a805c5f4d1be007cfcaef4b06fe363d402a3cc227b7c956f"} Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.027727 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-h7bw5" event={"ID":"26a37cc7-419b-489d-aa13-ec53651e212b","Type":"ContainerStarted","Data":"994a05b2da70ddeed736814a9dce29b53a5d938a179b8fb07353856cc08e67c6"} Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.030569 4810 generic.go:334] "Generic (PLEG): container finished" podID="d83a93fa-815c-4ec8-922e-c459c1fc0e7c" containerID="be58433decdd827b39ca07ace618938f8505204c8b8b85cbe4d49806b51713ed" exitCode=0 Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.030802 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" event={"ID":"d83a93fa-815c-4ec8-922e-c459c1fc0e7c","Type":"ContainerDied","Data":"be58433decdd827b39ca07ace618938f8505204c8b8b85cbe4d49806b51713ed"} Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.036144 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" event={"ID":"c8a4fee4-3426-4a48-8ae5-eec32d6d70ba","Type":"ContainerStarted","Data":"636487618f79984c2a7dc375d7e18fab61e217eadaad2b9d81852a241d45e927"} Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.038244 4810 patch_prober.go:28] interesting pod/console-operator-58897d9998-nhfv4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.038287 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" podUID="63a2b7dd-cd9b-44f3-b7bb-b37ae920a699" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.038935 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m85n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.039047 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6m85n" podUID="26503c29-9886-4ce9-b8f5-b8f03f1fa992" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.039105 4810 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-56cpr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.039136 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" podUID="4980c5d8-c614-4943-aff6-cb07dac7593f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.039144 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.048622 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.081722 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t8bxf"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.124716 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-752k2"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.125441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.126679 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.626659188 +0000 UTC m=+143.261098928 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.126966 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.188517 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.191953 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7q29k"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.195265 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.195312 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nqgd7"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.226917 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.227487 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.727473041 +0000 UTC m=+143.361912781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.329751 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.330269 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.830253718 +0000 UTC m=+143.464693458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.377142 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.431081 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.431704 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:00.931693353 +0000 UTC m=+143.566133093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.504540 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.532079 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.532501 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.032486675 +0000 UTC m=+143.666926415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.571232 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-jn2mz" podStartSLOduration=119.571219065 podStartE2EDuration="1m59.571219065s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:00.569438657 +0000 UTC m=+143.203878397" watchObservedRunningTime="2025-10-08 06:34:00.571219065 +0000 UTC m=+143.205658795" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.637162 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.637676 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.137659122 +0000 UTC m=+143.772098862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: W1008 06:34:00.667209 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12e1e52f_d7f4_4568_a0ac_79ce8102ff96.slice/crio-4ff7a79e7420cfc873de1c16b477c37617a4b253f9b4fa9535579f448cd5e2fe WatchSource:0}: Error finding container 4ff7a79e7420cfc873de1c16b477c37617a4b253f9b4fa9535579f448cd5e2fe: Status 404 returned error can't find the container with id 4ff7a79e7420cfc873de1c16b477c37617a4b253f9b4fa9535579f448cd5e2fe Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.685261 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.739641 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.739943 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.239929883 +0000 UTC m=+143.874369613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.758026 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.763671 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nptxs" podStartSLOduration=119.763644547 podStartE2EDuration="1m59.763644547s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:00.762113747 +0000 UTC m=+143.396553487" watchObservedRunningTime="2025-10-08 06:34:00.763644547 +0000 UTC m=+143.398084287" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.780212 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l"] Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.798748 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" podStartSLOduration=119.798724507 podStartE2EDuration="1m59.798724507s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:00.786174572 +0000 UTC m=+143.420614312" watchObservedRunningTime="2025-10-08 06:34:00.798724507 +0000 UTC m=+143.433164247" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.841200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.841564 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.341552133 +0000 UTC m=+143.975991873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.844782 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" podStartSLOduration=119.844766519 podStartE2EDuration="1m59.844766519s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:00.842516934 +0000 UTC m=+143.476956674" watchObservedRunningTime="2025-10-08 06:34:00.844766519 +0000 UTC m=+143.479206259" Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.862577 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hm95q"] Oct 08 06:34:00 crc kubenswrapper[4810]: W1008 06:34:00.924778 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa8545fa_ddc1_4af9_8095_f22465be5ed7.slice/crio-27811f95dec5f799b30b96a781aa15bdafed3331e06b432311a6fc979ba39c89 WatchSource:0}: Error finding container 27811f95dec5f799b30b96a781aa15bdafed3331e06b432311a6fc979ba39c89: Status 404 returned error can't find the container with id 27811f95dec5f799b30b96a781aa15bdafed3331e06b432311a6fc979ba39c89 Oct 08 06:34:00 crc kubenswrapper[4810]: I1008 06:34:00.956103 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:00 crc kubenswrapper[4810]: E1008 06:34:00.957253 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.457230097 +0000 UTC m=+144.091669837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.050279 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" podStartSLOduration=120.050261963 podStartE2EDuration="2m0.050261963s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.036424175 +0000 UTC m=+143.670863915" watchObservedRunningTime="2025-10-08 06:34:01.050261963 +0000 UTC m=+143.684701703" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.062263 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5wdms"] Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.065771 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.565748405 +0000 UTC m=+144.200188145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.082164 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.118366 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pmwfq"] Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.153806 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57"] Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.179241 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" event={"ID":"5651941c-e009-48b5-b824-69351bb54c6d","Type":"ContainerStarted","Data":"4156052a005e1c72ce5245e9c97043e22fecad18314d28a9c8f310569bfa25ce"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.183259 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.183569 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.68355423 +0000 UTC m=+144.317993970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.184349 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-h7bw5" event={"ID":"26a37cc7-419b-489d-aa13-ec53651e212b","Type":"ContainerStarted","Data":"695c8df8c4e36d6b30e43d01cbdc269c57c2b1a6f637e9296f17b8e921514098"} Oct 08 06:34:01 crc kubenswrapper[4810]: W1008 06:34:01.205719 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ba1679f_10be_4f8f_843b_0ca4bec2aaa5.slice/crio-926a70c6db70bab4f2cd511ec8d34d45d5e155e8821b5549dd921da4e8beba14 WatchSource:0}: Error finding container 926a70c6db70bab4f2cd511ec8d34d45d5e155e8821b5549dd921da4e8beba14: Status 404 returned error can't find the container with id 926a70c6db70bab4f2cd511ec8d34d45d5e155e8821b5549dd921da4e8beba14 Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.213216 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" event={"ID":"fa8545fa-ddc1-4af9-8095-f22465be5ed7","Type":"ContainerStarted","Data":"27811f95dec5f799b30b96a781aa15bdafed3331e06b432311a6fc979ba39c89"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.230611 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ch8bp" podStartSLOduration=120.230593965 podStartE2EDuration="2m0.230593965s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.229700235 +0000 UTC m=+143.864139975" watchObservedRunningTime="2025-10-08 06:34:01.230593965 +0000 UTC m=+143.865033705" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.249418 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6m85n" podStartSLOduration=120.249380296 podStartE2EDuration="2m0.249380296s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.245188358 +0000 UTC m=+143.879628098" watchObservedRunningTime="2025-10-08 06:34:01.249380296 +0000 UTC m=+143.883820036" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.287364 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.288870 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.788850682 +0000 UTC m=+144.423290422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.297499 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sqn9f" podStartSLOduration=120.297471587 podStartE2EDuration="2m0.297471587s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.288150018 +0000 UTC m=+143.922589758" watchObservedRunningTime="2025-10-08 06:34:01.297471587 +0000 UTC m=+143.931911327" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.315047 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" event={"ID":"472df6a2-664a-447c-9f9b-ad26fd30c300","Type":"ContainerStarted","Data":"12f3ca422e99a660a4189e6dfc0c18286e5580a24aa2a2cd5f565887cf7dc9cb"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.359587 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" event={"ID":"fda92a79-7c7e-4e9d-b67c-5ed1647946db","Type":"ContainerStarted","Data":"4ac15eb7fb5e412900cd2095edc6632327e4eee2f46ce197a4989e28968c10b2"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.363708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-752k2" event={"ID":"d1494115-1173-4a30-a886-bc2fa4fb97ab","Type":"ContainerStarted","Data":"a8ad02f26c94e1086606ec711811e36c62ae085e707066a7b29eddfd91c6bcfb"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.372400 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" event={"ID":"12e1e52f-d7f4-4568-a0ac-79ce8102ff96","Type":"ContainerStarted","Data":"4ff7a79e7420cfc873de1c16b477c37617a4b253f9b4fa9535579f448cd5e2fe"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.384909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" event={"ID":"bbe763ba-a910-4767-a673-7b7a3d76f342","Type":"ContainerStarted","Data":"fe1ab0f6500d93af3605dd50feab47076162e38fd5c0e4243029480a3886586a"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.385211 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" event={"ID":"bbe763ba-a910-4767-a673-7b7a3d76f342","Type":"ContainerStarted","Data":"82ef569cba4c187e10ea095a0bef532f1ab97378e79498051cde85c0c39bda2c"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.389730 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.390790 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.89074388 +0000 UTC m=+144.525183620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.407497 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" event={"ID":"ce854742-cd04-4ccc-81cd-895c46b3d566","Type":"ContainerStarted","Data":"e2e4341dc6dca5303042527c7af6e924379e72c22ef04c448be3d260fcacd6ac"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.421932 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" event={"ID":"74c3e024-0dfe-4f9c-930e-70896c3b8110","Type":"ContainerStarted","Data":"7c1e338e4bee426d2b94f80afe5f4d7a6e66e0f6ca18319d1f30bfc23f746ffc"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.423566 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" event={"ID":"ba1edb88-a5c3-4e50-8c6f-071554e22500","Type":"ContainerStarted","Data":"c00908d1313f9bbb8e8d1d617f17ddd893ff8558a05a45540af2d4e90d90f0ec"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.476442 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" podStartSLOduration=119.476412313 podStartE2EDuration="1m59.476412313s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.363261751 +0000 UTC m=+143.997701491" watchObservedRunningTime="2025-10-08 06:34:01.476412313 +0000 UTC m=+144.110852053" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.488037 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h"] Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.488259 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" event={"ID":"402ce4ba-4a2f-46c0-a810-be60006afd09","Type":"ContainerStarted","Data":"82e885522ad1a5bcacb4a40932aff90917c07f8758045ec650a7510bd37d1926"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.491874 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.492756 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:01.992741482 +0000 UTC m=+144.627181212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.498731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" event={"ID":"e2e82017-c4ed-4ac3-bfca-f117264ffaca","Type":"ContainerStarted","Data":"9a74e3da0d43842109bb5b99a3b228d5fa3394491f2fc00d0aa5cbc855a2efc5"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.506257 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" podStartSLOduration=120.506217728 podStartE2EDuration="2m0.506217728s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.486141694 +0000 UTC m=+144.120581434" watchObservedRunningTime="2025-10-08 06:34:01.506217728 +0000 UTC m=+144.140657468" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.508607 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" event={"ID":"2e808a56-2e57-410d-ace9-cd65a93881d6","Type":"ContainerStarted","Data":"cae9ec47e735ce4cccbfd42d36a99106202a74e6752ea87dc6f0873c95746a8e"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.553016 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" event={"ID":"c8a4fee4-3426-4a48-8ae5-eec32d6d70ba","Type":"ContainerStarted","Data":"b54ceb97a645e8140da94a774bb859e3d4e539f1da833d28cd7554a4b1b70d3b"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.593695 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.594817 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.094800996 +0000 UTC m=+144.729240736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: W1008 06:34:01.615057 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb125d88d_22fa_45cc_bfa0_564a6b9eb619.slice/crio-b5e67c916b3270b30d611a9075b1d6b8b2ccda3ce3a5ab1db5bf58e6837ec907 WatchSource:0}: Error finding container b5e67c916b3270b30d611a9075b1d6b8b2ccda3ce3a5ab1db5bf58e6837ec907: Status 404 returned error can't find the container with id b5e67c916b3270b30d611a9075b1d6b8b2ccda3ce3a5ab1db5bf58e6837ec907 Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.629284 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" event={"ID":"3868f3e7-62b6-4ba7-b4a7-69192183f6ed","Type":"ContainerStarted","Data":"cbb272c89d7c26e6649a7d3ce4fc39b074ee49a47a0066ba992c03c20be3677d"} Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.646935 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" podStartSLOduration=119.646914749 podStartE2EDuration="1m59.646914749s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.645141641 +0000 UTC m=+144.279581381" watchObservedRunningTime="2025-10-08 06:34:01.646914749 +0000 UTC m=+144.281354489" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.650941 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.682584 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lgrl2" podStartSLOduration=120.682566848 podStartE2EDuration="2m0.682566848s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.682387912 +0000 UTC m=+144.316827652" watchObservedRunningTime="2025-10-08 06:34:01.682566848 +0000 UTC m=+144.317006588" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.695320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.695671 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.195655401 +0000 UTC m=+144.830095151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.702156 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-nhfv4" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.723694 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-fq7wx" podStartSLOduration=120.723658447 podStartE2EDuration="2m0.723658447s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.719518549 +0000 UTC m=+144.353958289" watchObservedRunningTime="2025-10-08 06:34:01.723658447 +0000 UTC m=+144.358098187" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.763433 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-h7bw5" podStartSLOduration=5.763412431 podStartE2EDuration="5.763412431s" podCreationTimestamp="2025-10-08 06:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.754869389 +0000 UTC m=+144.389309129" watchObservedRunningTime="2025-10-08 06:34:01.763412431 +0000 UTC m=+144.397852171" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.804877 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.808590 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.308565654 +0000 UTC m=+144.943005394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.834307 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-987vk" podStartSLOduration=120.834272633 podStartE2EDuration="2m0.834272633s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.831345237 +0000 UTC m=+144.465784987" watchObservedRunningTime="2025-10-08 06:34:01.834272633 +0000 UTC m=+144.468712373" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.840095 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" podStartSLOduration=120.840080665 podStartE2EDuration="2m0.840080665s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:01.80333331 +0000 UTC m=+144.437773050" watchObservedRunningTime="2025-10-08 06:34:01.840080665 +0000 UTC m=+144.474520405" Oct 08 06:34:01 crc kubenswrapper[4810]: I1008 06:34:01.915790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:01 crc kubenswrapper[4810]: E1008 06:34:01.916351 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.416329087 +0000 UTC m=+145.050768827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.033626 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.034167 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.534146282 +0000 UTC m=+145.168586022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.136660 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.138047 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.638034947 +0000 UTC m=+145.272474677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.244623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.245240 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.745196419 +0000 UTC m=+145.379636159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.245695 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.246063 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.746051967 +0000 UTC m=+145.380491707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.347389 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.347774 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.84775574 +0000 UTC m=+145.482195480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.449250 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.449530 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:02.949518605 +0000 UTC m=+145.583958345 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.550845 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.551167 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.051137884 +0000 UTC m=+145.685577624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.551350 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.551716 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.051699482 +0000 UTC m=+145.686139222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.655790 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.656024 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.156002172 +0000 UTC m=+145.790441912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.656562 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.657913 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.157900694 +0000 UTC m=+145.792340424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.669355 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" event={"ID":"201eaacc-b3ab-4b74-9105-89d0ef91a74e","Type":"ContainerStarted","Data":"33d785e21a47e541d23a5f34c2e57266cafa09a7d3fdf5471dad24503a02c1a9"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.681044 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" event={"ID":"12e1e52f-d7f4-4568-a0ac-79ce8102ff96","Type":"ContainerStarted","Data":"6416023183ac13822e7d151967294d3d8aa002028e1bf1ad954594e7aff74f87"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.686710 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wcrhl" event={"ID":"3868f3e7-62b6-4ba7-b4a7-69192183f6ed","Type":"ContainerStarted","Data":"da3b2f4bd4b17329d0b304cf49b5ad7bfd985ab4fc097051d24b4bef5b3ed2a1"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.690889 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pmwfq" event={"ID":"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4","Type":"ContainerStarted","Data":"ce99d03211b7538147b639afd6c1b69e05d920d53f4d02e9c661327d47bfd1bc"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.695440 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" event={"ID":"b125d88d-22fa-45cc-bfa0-564a6b9eb619","Type":"ContainerStarted","Data":"b5e67c916b3270b30d611a9075b1d6b8b2ccda3ce3a5ab1db5bf58e6837ec907"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.716555 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-r68jt" podStartSLOduration=120.716540052 podStartE2EDuration="2m0.716540052s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.702347043 +0000 UTC m=+145.336786783" watchObservedRunningTime="2025-10-08 06:34:02.716540052 +0000 UTC m=+145.350979792" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.717794 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q75tf" podStartSLOduration=121.717786474 podStartE2EDuration="2m1.717786474s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.716869733 +0000 UTC m=+145.351309473" watchObservedRunningTime="2025-10-08 06:34:02.717786474 +0000 UTC m=+145.352226214" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.750581 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" event={"ID":"c8a4fee4-3426-4a48-8ae5-eec32d6d70ba","Type":"ContainerStarted","Data":"cdcec8f8557519702d9c90811cbe0ff967f31132c8a41b1cf5052951fe9ab22e"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.761325 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.762690 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.262675018 +0000 UTC m=+145.897114758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.771973 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-72rb2" podStartSLOduration=121.771953535 podStartE2EDuration="2m1.771953535s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.76908518 +0000 UTC m=+145.403524920" watchObservedRunningTime="2025-10-08 06:34:02.771953535 +0000 UTC m=+145.406393275" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.777436 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-fwdsd" event={"ID":"646d90cc-1f56-4fd1-bd7f-938021eeb92a","Type":"ContainerStarted","Data":"c86ebb34adffcea5bbc1782c91c79377cd18778998019255c231485b1733201c"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.783058 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" event={"ID":"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8","Type":"ContainerStarted","Data":"e7dacb40e93102e9a4899e95babd468adcb9baabdd43c02c3e9279ca7e72b08c"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.792848 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" event={"ID":"2f25e194-840d-4f74-a2e1-b0a81f4bb420","Type":"ContainerStarted","Data":"f656653f5b84e1450481e0ca6559d6f31203c0f1114e16f5f04a4813210b0383"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.803098 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" event={"ID":"86fe30b2-8748-4816-a44e-8428c396e209","Type":"ContainerStarted","Data":"3defd607588efeefd5c572f8d05cb62455d81a910a4e78cf184f4f5c4d704cf9"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.811395 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-fwdsd" podStartSLOduration=121.811374868 podStartE2EDuration="2m1.811374868s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.806177496 +0000 UTC m=+145.440617246" watchObservedRunningTime="2025-10-08 06:34:02.811374868 +0000 UTC m=+145.445814608" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.822419 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" event={"ID":"5651941c-e009-48b5-b824-69351bb54c6d","Type":"ContainerStarted","Data":"82e8dd0d5d945b34aa94556296ff3fbda59a6c32ea7f4d542ec806beae451103"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.824000 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" event={"ID":"53c6fda7-90ca-40b7-ba7d-2b182a090130","Type":"ContainerStarted","Data":"5fe72b281f4363b5898efa81b27485c762a2781bf502b781ea2179c544d264b3"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.834701 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" event={"ID":"472df6a2-664a-447c-9f9b-ad26fd30c300","Type":"ContainerStarted","Data":"4e375a736576cbb2251a7ce072c673e46158096d3863a4e38bb6324ffb1bd7d8"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.836640 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" event={"ID":"e2e82017-c4ed-4ac3-bfca-f117264ffaca","Type":"ContainerStarted","Data":"a55e764b0aba94520d55b472d47ee0008a73c79c0aaa0537e366229fe462abb8"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.837605 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" event={"ID":"3ee7a22a-3831-4afa-8093-74d09e3c0e4c","Type":"ContainerStarted","Data":"552e3463617c8af5c50c5029755e5735e369ce5797b1b0361ec5697a9931833c"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.845210 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.845368 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.847816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" event={"ID":"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5","Type":"ContainerStarted","Data":"926a70c6db70bab4f2cd511ec8d34d45d5e155e8821b5549dd921da4e8beba14"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.850272 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" event={"ID":"fda92a79-7c7e-4e9d-b67c-5ed1647946db","Type":"ContainerStarted","Data":"ed318e0609cf617ba66d82fa5e8408acee564d02ccd0aa54c5e30136d6f03c3a"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.860333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" event={"ID":"9d427010-0272-41f7-9883-4ef48df11e53","Type":"ContainerStarted","Data":"3b3fd433b5c436d38a65bc8f4c3f40bd7d3c357769a3aedbebd3bb6bca02d6f8"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.862155 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.862986 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.865801 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.365785696 +0000 UTC m=+146.000225436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.880415 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-752k2" event={"ID":"d1494115-1173-4a30-a886-bc2fa4fb97ab","Type":"ContainerStarted","Data":"a51cccc2991b0080dd458cd8e39f6e31d6dcc0cf30bcd32bd20e2ec85137c652"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.883839 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6dwqm" podStartSLOduration=120.883825343 podStartE2EDuration="2m0.883825343s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.85681196 +0000 UTC m=+145.491251700" watchObservedRunningTime="2025-10-08 06:34:02.883825343 +0000 UTC m=+145.518265083" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.893150 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" event={"ID":"74c3e024-0dfe-4f9c-930e-70896c3b8110","Type":"ContainerStarted","Data":"a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d"} Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.902697 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-whsfl" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.911615 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6x4v5" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.932745 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2jfkb" podStartSLOduration=121.932719909 podStartE2EDuration="2m1.932719909s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.889947306 +0000 UTC m=+145.524387046" watchObservedRunningTime="2025-10-08 06:34:02.932719909 +0000 UTC m=+145.567159649" Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.963789 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-m89fc" podStartSLOduration=121.963773186 podStartE2EDuration="2m1.963773186s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.933232317 +0000 UTC m=+145.567672057" watchObservedRunningTime="2025-10-08 06:34:02.963773186 +0000 UTC m=+145.598212916" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.963884 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.463864349 +0000 UTC m=+146.098304089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.963805 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:02 crc kubenswrapper[4810]: I1008 06:34:02.964372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:02 crc kubenswrapper[4810]: E1008 06:34:02.966039 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.466027871 +0000 UTC m=+146.100467711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.016452 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" podStartSLOduration=121.016435087 podStartE2EDuration="2m1.016435087s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:02.97355513 +0000 UTC m=+145.607994870" watchObservedRunningTime="2025-10-08 06:34:03.016435087 +0000 UTC m=+145.650874827" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.089624 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.089788 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.589763322 +0000 UTC m=+146.224203062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.090082 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.090519 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.590507387 +0000 UTC m=+146.224947127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.194078 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.194233 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.694213046 +0000 UTC m=+146.328652786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.194585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.194959 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.694943099 +0000 UTC m=+146.329382839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.296299 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.296470 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.796450945 +0000 UTC m=+146.430890685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.397944 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.398419 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.898401736 +0000 UTC m=+146.532841476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.499042 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.499322 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.999272321 +0000 UTC m=+146.633712081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.499438 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.499901 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:03.999887971 +0000 UTC m=+146.634327711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.600444 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.600625 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.100588801 +0000 UTC m=+146.735028541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.600896 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.601342 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.101327915 +0000 UTC m=+146.735767655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.668800 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.679018 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:03 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:03 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:03 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.679089 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.701882 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.702126 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.202095246 +0000 UTC m=+146.836534986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.702245 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.702733 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.202715497 +0000 UTC m=+146.837155237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.803563 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.803763 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.303732836 +0000 UTC m=+146.938172576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.804191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.804619 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.304603856 +0000 UTC m=+146.939043596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.899537 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" event={"ID":"d83a93fa-815c-4ec8-922e-c459c1fc0e7c","Type":"ContainerStarted","Data":"2be9c081cbe80a63226f939ca612995f1ed432f81b30e9a7585727e8e819d76b"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.899985 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" event={"ID":"d83a93fa-815c-4ec8-922e-c459c1fc0e7c","Type":"ContainerStarted","Data":"72aa5481852586a23395d5e47a22f9842e5f752c01a297457aa2bf500dd29ce4"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.900693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" event={"ID":"315ec44f-4c73-4ce8-afaa-4a0ec9b37cd8","Type":"ContainerStarted","Data":"b5bf036905fd2c8b11f18ce0665e0e430cfdd85b2d45985442b3e6b8a87ad911"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.903593 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" event={"ID":"402ce4ba-4a2f-46c0-a810-be60006afd09","Type":"ContainerStarted","Data":"018caf9a3bb3774f5b5eb2d5056bb2c6363bb58a89637ddd69f58fdde183c974"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.903703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" event={"ID":"402ce4ba-4a2f-46c0-a810-be60006afd09","Type":"ContainerStarted","Data":"d2d3e2268475b26803e3bf638e922eef937050c2c98c29e6b12b3e9ddf5e169f"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.904824 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.905005 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.404979503 +0000 UTC m=+147.039419243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.905431 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:03 crc kubenswrapper[4810]: E1008 06:34:03.906598 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.406585227 +0000 UTC m=+147.041024957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.907147 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" event={"ID":"fa8545fa-ddc1-4af9-8095-f22465be5ed7","Type":"ContainerStarted","Data":"81d5dcbfa9274f5db92a04b8f0651572e951bc2cdcd11c05603c0c5c68406a88"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.907194 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" event={"ID":"fa8545fa-ddc1-4af9-8095-f22465be5ed7","Type":"ContainerStarted","Data":"d827fdac315863cbf05cd272d020d47fa58501bf2044d9098f826ccdd3ccb221"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.907303 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.909577 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" event={"ID":"e2e82017-c4ed-4ac3-bfca-f117264ffaca","Type":"ContainerStarted","Data":"9f2f4bc5c8c9aab22fe4f7a64a038120e519b0f468ba90c6e247bd6ba243122f"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.911379 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pmwfq" event={"ID":"701ecb4f-1521-41ce-8294-3e2ec4a1f1a4","Type":"ContainerStarted","Data":"e5f48813508162906f1fc9fb87f7f0958a79b0c488687f051701a3e945f355b0"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.913760 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" event={"ID":"9d427010-0272-41f7-9883-4ef48df11e53","Type":"ContainerStarted","Data":"121e0f07d8aa4a267201ca3280dbd175585b212e1ef4af1b9756af2fc210093c"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.915579 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" event={"ID":"b125d88d-22fa-45cc-bfa0-564a6b9eb619","Type":"ContainerStarted","Data":"0b575f7fa9529bb4ca23cd89ca3e7d6622d5a1006aa426cb7d464b512fac597b"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.915682 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" event={"ID":"b125d88d-22fa-45cc-bfa0-564a6b9eb619","Type":"ContainerStarted","Data":"8874e4833e34ef6e3a585b126de375fa3e961a239e370411ff9bc584dd93bd85"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.917379 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" event={"ID":"bbe763ba-a910-4767-a673-7b7a3d76f342","Type":"ContainerStarted","Data":"eef26abecfeee1014293795d52f7c01cde6d1facb1c81088526d356b103b80ee"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.919346 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-752k2" event={"ID":"d1494115-1173-4a30-a886-bc2fa4fb97ab","Type":"ContainerStarted","Data":"cec06eed9d71c3bd3ee0eb81637532b239e13bc0289cd4b8fce8a86429b36918"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.919743 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-752k2" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.921153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" event={"ID":"2ba1679f-10be-4f8f-843b-0ca4bec2aaa5","Type":"ContainerStarted","Data":"dca136b793cceecda6b9a78274f0cbede72e35566d7bd7dfd6ad935050d41a4d"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.921600 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.923447 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" event={"ID":"2e808a56-2e57-410d-ace9-cd65a93881d6","Type":"ContainerStarted","Data":"bf09f80754ef5c1c514f1533bd8c663941877611a6799b4174b51bbae2d2fa3f"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.924000 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.924596 4810 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-q9sq8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.924798 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" podUID="2ba1679f-10be-4f8f-843b-0ca4bec2aaa5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.925156 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" event={"ID":"86fe30b2-8748-4816-a44e-8428c396e209","Type":"ContainerStarted","Data":"9b313b25d21003958bdce8b33077218c6892c64eea109fa7993ef5d70574d03e"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.925709 4810 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-wfshr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.925771 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" podUID="2e808a56-2e57-410d-ace9-cd65a93881d6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.926589 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" event={"ID":"2f25e194-840d-4f74-a2e1-b0a81f4bb420","Type":"ContainerStarted","Data":"ae419ea53a1f94b9977f7de8d74dce3c342dd0282d605d8cf59b793bbd14e60d"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.926924 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.928346 4810 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-2xp4l container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.928376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" event={"ID":"ba1edb88-a5c3-4e50-8c6f-071554e22500","Type":"ContainerStarted","Data":"97aec8f876487ba51e349fbbfdb8df7044792b85b2a65133c187b78ad0c6a441"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.928437 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" podUID="2f25e194-840d-4f74-a2e1-b0a81f4bb420" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.929782 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" event={"ID":"53c6fda7-90ca-40b7-ba7d-2b182a090130","Type":"ContainerStarted","Data":"82368bea6e45f23904dc6d0167b9b667d7495df2962ee634c6fcaa96164c38bb"} Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.930699 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.932792 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7q29k container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.932834 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" podUID="74c3e024-0dfe-4f9c-930e-70896c3b8110" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.951059 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" podStartSLOduration=122.950800889 podStartE2EDuration="2m2.950800889s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:03.947671056 +0000 UTC m=+146.582110796" watchObservedRunningTime="2025-10-08 06:34:03.950800889 +0000 UTC m=+146.585240629" Oct 08 06:34:03 crc kubenswrapper[4810]: I1008 06:34:03.971249 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5wdms" podStartSLOduration=121.971229034 podStartE2EDuration="2m1.971229034s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:03.970615254 +0000 UTC m=+146.605054994" watchObservedRunningTime="2025-10-08 06:34:03.971229034 +0000 UTC m=+146.605668774" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.007373 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.008136 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" podStartSLOduration=123.008119814 podStartE2EDuration="2m3.008119814s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.005412964 +0000 UTC m=+146.639852704" watchObservedRunningTime="2025-10-08 06:34:04.008119814 +0000 UTC m=+146.642559554" Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.009648 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.509615123 +0000 UTC m=+147.144054863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.040282 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hm95q" podStartSLOduration=123.040252877 podStartE2EDuration="2m3.040252877s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.039423398 +0000 UTC m=+146.673863138" watchObservedRunningTime="2025-10-08 06:34:04.040252877 +0000 UTC m=+146.674692617" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.116939 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.117106 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-752k2" podStartSLOduration=8.117076576 podStartE2EDuration="8.117076576s" podCreationTimestamp="2025-10-08 06:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.100307562 +0000 UTC m=+146.734747302" watchObservedRunningTime="2025-10-08 06:34:04.117076576 +0000 UTC m=+146.751516306" Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.117342 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.617326684 +0000 UTC m=+147.251766424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.117969 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" podStartSLOduration=122.117957455 podStartE2EDuration="2m2.117957455s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.074155647 +0000 UTC m=+146.708595387" watchObservedRunningTime="2025-10-08 06:34:04.117957455 +0000 UTC m=+146.752397195" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.157714 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-f779z" podStartSLOduration=122.157696749 podStartE2EDuration="2m2.157696749s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.155169276 +0000 UTC m=+146.789609016" watchObservedRunningTime="2025-10-08 06:34:04.157696749 +0000 UTC m=+146.792136489" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.217521 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.217845 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.717829977 +0000 UTC m=+147.352269717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.243533 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pmwfq" podStartSLOduration=8.243515696 podStartE2EDuration="8.243515696s" podCreationTimestamp="2025-10-08 06:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.200704041 +0000 UTC m=+146.835143781" watchObservedRunningTime="2025-10-08 06:34:04.243515696 +0000 UTC m=+146.877955436" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.244447 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5jc7h" podStartSLOduration=123.244438707 podStartE2EDuration="2m3.244438707s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.242432021 +0000 UTC m=+146.876871771" watchObservedRunningTime="2025-10-08 06:34:04.244438707 +0000 UTC m=+146.878878457" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.270435 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" podStartSLOduration=122.270417716 podStartE2EDuration="2m2.270417716s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.268111799 +0000 UTC m=+146.902551539" watchObservedRunningTime="2025-10-08 06:34:04.270417716 +0000 UTC m=+146.904857456" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.306497 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" podStartSLOduration=122.306480488 podStartE2EDuration="2m2.306480488s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.298256236 +0000 UTC m=+146.932695966" watchObservedRunningTime="2025-10-08 06:34:04.306480488 +0000 UTC m=+146.940920228" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.320401 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.320759 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.82074728 +0000 UTC m=+147.455187020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.371895 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q6wf9" podStartSLOduration=123.37187977 podStartE2EDuration="2m3.37187977s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.351443145 +0000 UTC m=+146.985882885" watchObservedRunningTime="2025-10-08 06:34:04.37187977 +0000 UTC m=+147.006319510" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.394705 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" podStartSLOduration=122.394693794 podStartE2EDuration="2m2.394693794s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.392406828 +0000 UTC m=+147.026846568" watchObservedRunningTime="2025-10-08 06:34:04.394693794 +0000 UTC m=+147.029133534" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.394799 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rx5cd" podStartSLOduration=122.394794918 podStartE2EDuration="2m2.394794918s" podCreationTimestamp="2025-10-08 06:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.372139929 +0000 UTC m=+147.006579669" watchObservedRunningTime="2025-10-08 06:34:04.394794918 +0000 UTC m=+147.029234658" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.413525 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-nqgd7" podStartSLOduration=123.413508857 podStartE2EDuration="2m3.413508857s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.412372599 +0000 UTC m=+147.046812339" watchObservedRunningTime="2025-10-08 06:34:04.413508857 +0000 UTC m=+147.047948587" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.421405 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.421544 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.921510661 +0000 UTC m=+147.555950401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.421604 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.421925 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:04.921913275 +0000 UTC m=+147.556353005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.437752 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z7t57" podStartSLOduration=123.437737128 podStartE2EDuration="2m3.437737128s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:04.437266502 +0000 UTC m=+147.071706242" watchObservedRunningTime="2025-10-08 06:34:04.437737128 +0000 UTC m=+147.072176868" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.522459 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.522592 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.022569592 +0000 UTC m=+147.657009332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.522972 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.523326 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.023315157 +0000 UTC m=+147.657754897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.623643 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.623933 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.123919683 +0000 UTC m=+147.758359423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.679490 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:04 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:04 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:04 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.679558 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.724779 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.725158 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.22514754 +0000 UTC m=+147.859587270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.825782 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.826063 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.326037455 +0000 UTC m=+147.960477195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.826198 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.826473 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.326460149 +0000 UTC m=+147.960899889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.926965 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:04 crc kubenswrapper[4810]: E1008 06:34:04.927367 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.427349144 +0000 UTC m=+148.061788884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.952428 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7q29k container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.952485 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" podUID="74c3e024-0dfe-4f9c-930e-70896c3b8110" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 08 06:34:04 crc kubenswrapper[4810]: I1008 06:34:04.964637 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2xp4l" Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.031258 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.032198 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.532184241 +0000 UTC m=+148.166623981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.041407 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wfshr" Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.132528 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.132713 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.632687174 +0000 UTC m=+148.267126914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.133006 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.136662 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.636648154 +0000 UTC m=+148.271087894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.235346 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.235523 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.735496802 +0000 UTC m=+148.369936542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.235725 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.236106 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.736092062 +0000 UTC m=+148.370531812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.337068 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.337302 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.837270537 +0000 UTC m=+148.471710287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.337359 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.337681 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.83765728 +0000 UTC m=+148.472097100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.437917 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.438131 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.938105601 +0000 UTC m=+148.572545341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.438248 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.438531 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:05.938518955 +0000 UTC m=+148.572958695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.539157 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.539476 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.039462242 +0000 UTC m=+148.673901982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.614660 4810 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.640960 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.641511 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.141494585 +0000 UTC m=+148.775934325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.673459 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:05 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:05 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:05 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.673543 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.678317 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-q9sq8" Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.742175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.742441 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.242399201 +0000 UTC m=+148.876838941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.743091 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.743669 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.243655313 +0000 UTC m=+148.878095053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.844149 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.844619 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.34457956 +0000 UTC m=+148.979019300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.945876 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:05 crc kubenswrapper[4810]: E1008 06:34:05.946258 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.446241291 +0000 UTC m=+149.080681031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.949248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" event={"ID":"ba1edb88-a5c3-4e50-8c6f-071554e22500","Type":"ContainerStarted","Data":"8325c243f37d274c9adfcb4d04191b7f5378e8820d7d6939138c9c3e89829a0e"} Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.949401 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" event={"ID":"ba1edb88-a5c3-4e50-8c6f-071554e22500","Type":"ContainerStarted","Data":"e87520429a4fb610836352d60196b4f172892727240de8d1ab333ae16e81fe5a"} Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.949483 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" event={"ID":"ba1edb88-a5c3-4e50-8c6f-071554e22500","Type":"ContainerStarted","Data":"dbe062f39cd00e3d11fcf8d4f27c269761c1b0b3edbf88bb5b12578f7e42617f"} Oct 08 06:34:05 crc kubenswrapper[4810]: I1008 06:34:05.976399 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-t8bxf" podStartSLOduration=9.976372837 podStartE2EDuration="9.976372837s" podCreationTimestamp="2025-10-08 06:33:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:05.975580091 +0000 UTC m=+148.610019851" watchObservedRunningTime="2025-10-08 06:34:05.976372837 +0000 UTC m=+148.610812577" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.046421 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.046572 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:34:06 crc kubenswrapper[4810]: E1008 06:34:06.046708 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.546664851 +0000 UTC m=+149.181104591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.046985 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.047113 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.047919 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:34:06 crc kubenswrapper[4810]: E1008 06:34:06.048629 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.548611435 +0000 UTC m=+149.183051365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.057729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.099823 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.148615 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.148971 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:34:06 crc kubenswrapper[4810]: E1008 06:34:06.149220 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.649190811 +0000 UTC m=+149.283630551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.149373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.157614 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.158439 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.251139 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:06 crc kubenswrapper[4810]: E1008 06:34:06.251937 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 06:34:06.751909256 +0000 UTC m=+149.386349176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gpgrc" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.296054 4810 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-08T06:34:05.615095993Z","Handler":null,"Name":""} Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.302455 4810 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.302504 4810 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.321858 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kb4ww"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.322860 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.326014 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.339755 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kb4ww"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.352737 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.352892 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzlmw\" (UniqueName: \"kubernetes.io/projected/85ab7796-c612-4b27-a367-1d3deca20d7a-kube-api-access-bzlmw\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.352920 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-utilities\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.352968 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-catalog-content\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.366274 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.414047 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.426099 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 06:34:06 crc kubenswrapper[4810]: W1008 06:34:06.446708 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-86c6ad43422ffb1c63e1ffdc98e7a5a79c8ecdf9b039ad4bb3ff24084f93cd57 WatchSource:0}: Error finding container 86c6ad43422ffb1c63e1ffdc98e7a5a79c8ecdf9b039ad4bb3ff24084f93cd57: Status 404 returned error can't find the container with id 86c6ad43422ffb1c63e1ffdc98e7a5a79c8ecdf9b039ad4bb3ff24084f93cd57 Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.453998 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-utilities\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.454115 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.454149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-catalog-content\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.454254 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzlmw\" (UniqueName: \"kubernetes.io/projected/85ab7796-c612-4b27-a367-1d3deca20d7a-kube-api-access-bzlmw\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.454559 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-utilities\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.455623 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-catalog-content\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.460550 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.460612 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.473732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzlmw\" (UniqueName: \"kubernetes.io/projected/85ab7796-c612-4b27-a367-1d3deca20d7a-kube-api-access-bzlmw\") pod \"community-operators-kb4ww\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.487996 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gpgrc\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.511915 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7fw7n"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.513365 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.517013 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.522165 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7fw7n"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.526052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.557834 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pccb4\" (UniqueName: \"kubernetes.io/projected/3abe3edb-3d3f-48a0-9968-f47012715666-kube-api-access-pccb4\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.557889 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-catalog-content\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.557941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-utilities\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.657372 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.660960 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-utilities\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.661101 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pccb4\" (UniqueName: \"kubernetes.io/projected/3abe3edb-3d3f-48a0-9968-f47012715666-kube-api-access-pccb4\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.661160 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-catalog-content\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.661789 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-catalog-content\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.662104 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-utilities\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.676506 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:06 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:06 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:06 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.676613 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.695370 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pccb4\" (UniqueName: \"kubernetes.io/projected/3abe3edb-3d3f-48a0-9968-f47012715666-kube-api-access-pccb4\") pod \"certified-operators-7fw7n\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.711678 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l98x5"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.712890 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.743803 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l98x5"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.761320 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.762134 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.763415 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-catalog-content\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.763448 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-utilities\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.763468 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwpf\" (UniqueName: \"kubernetes.io/projected/559f47bc-728a-4ef9-bff7-bcda90c4134e-kube-api-access-jgwpf\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.767313 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.767495 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.786885 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.848515 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.864479 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.864522 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-catalog-content\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.864555 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-utilities\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.864572 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.864592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgwpf\" (UniqueName: \"kubernetes.io/projected/559f47bc-728a-4ef9-bff7-bcda90c4134e-kube-api-access-jgwpf\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.865021 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-utilities\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.865201 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-catalog-content\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.889502 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgwpf\" (UniqueName: \"kubernetes.io/projected/559f47bc-728a-4ef9-bff7-bcda90c4134e-kube-api-access-jgwpf\") pod \"community-operators-l98x5\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.913300 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bhgcd"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.914674 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.951418 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhgcd"] Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.966010 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vm9j\" (UniqueName: \"kubernetes.io/projected/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-kube-api-access-6vm9j\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.966062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.966095 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-catalog-content\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.966136 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.966202 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-utilities\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.966428 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.967254 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c3330537b45026cb1e89d81ff6472a8600a9de5b901f12d16d6e943319082db0"} Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.988266 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4090a64f5cfd1b4c6be6c6acf9b73ca58391061e7e46adeadaf1902924f54f4f"} Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.989604 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2aba332718db4d17c9d47cb56876ff5a99ead4f9d9d4897d2d4d4dbfa4e2da54"} Oct 08 06:34:06 crc kubenswrapper[4810]: I1008 06:34:06.989648 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"86c6ad43422ffb1c63e1ffdc98e7a5a79c8ecdf9b039ad4bb3ff24084f93cd57"} Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.002924 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.017039 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gpgrc"] Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.061613 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.068807 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-utilities\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.068878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vm9j\" (UniqueName: \"kubernetes.io/projected/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-kube-api-access-6vm9j\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.068904 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-catalog-content\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.070376 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-utilities\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.071564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-catalog-content\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.089171 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.093211 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vm9j\" (UniqueName: \"kubernetes.io/projected/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-kube-api-access-6vm9j\") pod \"certified-operators-bhgcd\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.099186 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kb4ww"] Oct 08 06:34:07 crc kubenswrapper[4810]: W1008 06:34:07.141118 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85ab7796_c612_4b27_a367_1d3deca20d7a.slice/crio-6e52cbe58263d14b93494349105f4c2d1cd5861513fb8137cfe14e13f5d810ac WatchSource:0}: Error finding container 6e52cbe58263d14b93494349105f4c2d1cd5861513fb8137cfe14e13f5d810ac: Status 404 returned error can't find the container with id 6e52cbe58263d14b93494349105f4c2d1cd5861513fb8137cfe14e13f5d810ac Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.282426 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7fw7n"] Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.287001 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:07 crc kubenswrapper[4810]: W1008 06:34:07.323237 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3abe3edb_3d3f_48a0_9968_f47012715666.slice/crio-a8fb24019003f83f60f9f25fb3f2385cacad6f18c99ffe1d243c16542a9ae411 WatchSource:0}: Error finding container a8fb24019003f83f60f9f25fb3f2385cacad6f18c99ffe1d243c16542a9ae411: Status 404 returned error can't find the container with id a8fb24019003f83f60f9f25fb3f2385cacad6f18c99ffe1d243c16542a9ae411 Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.372713 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l98x5"] Oct 08 06:34:07 crc kubenswrapper[4810]: W1008 06:34:07.439899 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod559f47bc_728a_4ef9_bff7_bcda90c4134e.slice/crio-0053b534a76f7d8d5f287cbe7c7df7965d77a0260637956ec06d975f4b3aafa8 WatchSource:0}: Error finding container 0053b534a76f7d8d5f287cbe7c7df7965d77a0260637956ec06d975f4b3aafa8: Status 404 returned error can't find the container with id 0053b534a76f7d8d5f287cbe7c7df7965d77a0260637956ec06d975f4b3aafa8 Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.488704 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.586654 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhgcd"] Oct 08 06:34:07 crc kubenswrapper[4810]: W1008 06:34:07.660510 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73f9f84f_5f3c_45fa_8b5f_108bd77b2251.slice/crio-608a272c8f1572a5bbeff3b2e75fdbc0b9de2dab32cf28dba12f5a3b2e3fde79 WatchSource:0}: Error finding container 608a272c8f1572a5bbeff3b2e75fdbc0b9de2dab32cf28dba12f5a3b2e3fde79: Status 404 returned error can't find the container with id 608a272c8f1572a5bbeff3b2e75fdbc0b9de2dab32cf28dba12f5a3b2e3fde79 Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.673033 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:07 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:07 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:07 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.673100 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.705240 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.705304 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.711173 4810 patch_prober.go:28] interesting pod/console-f9d7485db-fq7wx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.711285 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fq7wx" podUID="303e572d-0bdf-4cf2-85bf-976e5ac59130" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.997038 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m85n container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.997350 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5","Type":"ContainerStarted","Data":"659959862fd18bb04954b8cb571616e1e803bea3801ed607f8e0687f388d8042"} Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.997391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5","Type":"ContainerStarted","Data":"717a550925a5bc76fec0f878359b8e0a39c50959a6635b8399577613a05b2645"} Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.997038 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m85n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.997395 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6m85n" podUID="26503c29-9886-4ce9-b8f5-b8f03f1fa992" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.997430 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6m85n" podUID="26503c29-9886-4ce9-b8f5-b8f03f1fa992" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.999279 4810 generic.go:334] "Generic (PLEG): container finished" podID="3abe3edb-3d3f-48a0-9968-f47012715666" containerID="cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e" exitCode=0 Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.999349 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fw7n" event={"ID":"3abe3edb-3d3f-48a0-9968-f47012715666","Type":"ContainerDied","Data":"cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e"} Oct 08 06:34:07 crc kubenswrapper[4810]: I1008 06:34:07.999376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fw7n" event={"ID":"3abe3edb-3d3f-48a0-9968-f47012715666","Type":"ContainerStarted","Data":"a8fb24019003f83f60f9f25fb3f2385cacad6f18c99ffe1d243c16542a9ae411"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.001088 4810 generic.go:334] "Generic (PLEG): container finished" podID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerID="151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c" exitCode=0 Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.001279 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l98x5" event={"ID":"559f47bc-728a-4ef9-bff7-bcda90c4134e","Type":"ContainerDied","Data":"151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.001395 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l98x5" event={"ID":"559f47bc-728a-4ef9-bff7-bcda90c4134e","Type":"ContainerStarted","Data":"0053b534a76f7d8d5f287cbe7c7df7965d77a0260637956ec06d975f4b3aafa8"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.001147 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.003538 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"08ca7c1562eba02c4b778094b416c4c592a09b42c043b50e97b9e41f20fcd534"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.008507 4810 generic.go:334] "Generic (PLEG): container finished" podID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerID="a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e" exitCode=0 Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.010209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb4ww" event={"ID":"85ab7796-c612-4b27-a367-1d3deca20d7a","Type":"ContainerDied","Data":"a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.010330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb4ww" event={"ID":"85ab7796-c612-4b27-a367-1d3deca20d7a","Type":"ContainerStarted","Data":"6e52cbe58263d14b93494349105f4c2d1cd5861513fb8137cfe14e13f5d810ac"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.012662 4810 generic.go:334] "Generic (PLEG): container finished" podID="5651941c-e009-48b5-b824-69351bb54c6d" containerID="82e8dd0d5d945b34aa94556296ff3fbda59a6c32ea7f4d542ec806beae451103" exitCode=0 Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.012718 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" event={"ID":"5651941c-e009-48b5-b824-69351bb54c6d","Type":"ContainerDied","Data":"82e8dd0d5d945b34aa94556296ff3fbda59a6c32ea7f4d542ec806beae451103"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.015964 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" event={"ID":"a714d476-65eb-4e18-a0ba-e874bae2ca79","Type":"ContainerStarted","Data":"2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.016115 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" event={"ID":"a714d476-65eb-4e18-a0ba-e874bae2ca79","Type":"ContainerStarted","Data":"4e1597d5658f0563aedccc5e43ed2447e78dbf8639a010a134597380cc78a79d"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.016705 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.019368 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"91ea441efce25a5a21aa6a742272f13721e1c4e8855cab71bb0393b17ff81157"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.019491 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.020920 4810 generic.go:334] "Generic (PLEG): container finished" podID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerID="eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5" exitCode=0 Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.020967 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhgcd" event={"ID":"73f9f84f-5f3c-45fa-8b5f-108bd77b2251","Type":"ContainerDied","Data":"eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.021010 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhgcd" event={"ID":"73f9f84f-5f3c-45fa-8b5f-108bd77b2251","Type":"ContainerStarted","Data":"608a272c8f1572a5bbeff3b2e75fdbc0b9de2dab32cf28dba12f5a3b2e3fde79"} Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.025878 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.025913 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.040916 4810 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kt2pp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]log ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]etcd ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/generic-apiserver-start-informers ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/max-in-flight-filter ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 08 06:34:08 crc kubenswrapper[4810]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/project.openshift.io-projectcache ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/openshift.io-startinformers ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 08 06:34:08 crc kubenswrapper[4810]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 08 06:34:08 crc kubenswrapper[4810]: livez check failed Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.041773 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" podUID="d83a93fa-815c-4ec8-922e-c459c1fc0e7c" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.044130 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.044112769 podStartE2EDuration="2.044112769s" podCreationTimestamp="2025-10-08 06:34:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:08.022027849 +0000 UTC m=+150.656467589" watchObservedRunningTime="2025-10-08 06:34:08.044112769 +0000 UTC m=+150.678552499" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.044922 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" podStartSLOduration=127.044916895 podStartE2EDuration="2m7.044916895s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:08.041897916 +0000 UTC m=+150.676337656" watchObservedRunningTime="2025-10-08 06:34:08.044916895 +0000 UTC m=+150.679356635" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.097503 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.508492 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xtjk2"] Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.509450 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.511670 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.528841 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xtjk2"] Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.603218 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-utilities\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.603258 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6b8b\" (UniqueName: \"kubernetes.io/projected/432c1370-925d-4d16-9bc0-b7736387e058-kube-api-access-b6b8b\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.603285 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-catalog-content\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.673761 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:08 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:08 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:08 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.673860 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.711210 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-utilities\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.711299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6b8b\" (UniqueName: \"kubernetes.io/projected/432c1370-925d-4d16-9bc0-b7736387e058-kube-api-access-b6b8b\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.711336 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-catalog-content\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.711953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-catalog-content\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.712237 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-utilities\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.748244 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6b8b\" (UniqueName: \"kubernetes.io/projected/432c1370-925d-4d16-9bc0-b7736387e058-kube-api-access-b6b8b\") pod \"redhat-marketplace-xtjk2\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.879089 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.910654 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xh74p"] Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.912213 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:08 crc kubenswrapper[4810]: I1008 06:34:08.921820 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh74p"] Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.019999 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42pz5\" (UniqueName: \"kubernetes.io/projected/87bfa11d-2272-489c-8a91-2b0e22acf8dd-kube-api-access-42pz5\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.020052 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-utilities\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.020097 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-catalog-content\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.034995 4810 generic.go:334] "Generic (PLEG): container finished" podID="84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5" containerID="659959862fd18bb04954b8cb571616e1e803bea3801ed607f8e0687f388d8042" exitCode=0 Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.035116 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5","Type":"ContainerDied","Data":"659959862fd18bb04954b8cb571616e1e803bea3801ed607f8e0687f388d8042"} Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.121282 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42pz5\" (UniqueName: \"kubernetes.io/projected/87bfa11d-2272-489c-8a91-2b0e22acf8dd-kube-api-access-42pz5\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.121359 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-utilities\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.121465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-catalog-content\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.122459 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-catalog-content\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.122742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-utilities\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.142870 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42pz5\" (UniqueName: \"kubernetes.io/projected/87bfa11d-2272-489c-8a91-2b0e22acf8dd-kube-api-access-42pz5\") pod \"redhat-marketplace-xh74p\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.249478 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xtjk2"] Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.287628 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.405456 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.432262 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5651941c-e009-48b5-b824-69351bb54c6d-secret-volume\") pod \"5651941c-e009-48b5-b824-69351bb54c6d\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.432383 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bg2h\" (UniqueName: \"kubernetes.io/projected/5651941c-e009-48b5-b824-69351bb54c6d-kube-api-access-5bg2h\") pod \"5651941c-e009-48b5-b824-69351bb54c6d\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.432420 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5651941c-e009-48b5-b824-69351bb54c6d-config-volume\") pod \"5651941c-e009-48b5-b824-69351bb54c6d\" (UID: \"5651941c-e009-48b5-b824-69351bb54c6d\") " Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.433519 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5651941c-e009-48b5-b824-69351bb54c6d-config-volume" (OuterVolumeSpecName: "config-volume") pod "5651941c-e009-48b5-b824-69351bb54c6d" (UID: "5651941c-e009-48b5-b824-69351bb54c6d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.440187 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5651941c-e009-48b5-b824-69351bb54c6d-kube-api-access-5bg2h" (OuterVolumeSpecName: "kube-api-access-5bg2h") pod "5651941c-e009-48b5-b824-69351bb54c6d" (UID: "5651941c-e009-48b5-b824-69351bb54c6d"). InnerVolumeSpecName "kube-api-access-5bg2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.442462 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5651941c-e009-48b5-b824-69351bb54c6d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5651941c-e009-48b5-b824-69351bb54c6d" (UID: "5651941c-e009-48b5-b824-69351bb54c6d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.510588 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8xxs"] Oct 08 06:34:09 crc kubenswrapper[4810]: E1008 06:34:09.512605 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5651941c-e009-48b5-b824-69351bb54c6d" containerName="collect-profiles" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.512620 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5651941c-e009-48b5-b824-69351bb54c6d" containerName="collect-profiles" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.514368 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5651941c-e009-48b5-b824-69351bb54c6d" containerName="collect-profiles" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.516413 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.519897 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.525703 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8xxs"] Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.534631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-utilities\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.534711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6ck7\" (UniqueName: \"kubernetes.io/projected/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-kube-api-access-r6ck7\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.534745 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-catalog-content\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.534854 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5651941c-e009-48b5-b824-69351bb54c6d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.534869 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bg2h\" (UniqueName: \"kubernetes.io/projected/5651941c-e009-48b5-b824-69351bb54c6d-kube-api-access-5bg2h\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.534880 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5651941c-e009-48b5-b824-69351bb54c6d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.636966 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6ck7\" (UniqueName: \"kubernetes.io/projected/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-kube-api-access-r6ck7\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.637054 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-catalog-content\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.637131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-utilities\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.637571 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-catalog-content\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.637762 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-utilities\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.662043 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.662387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6ck7\" (UniqueName: \"kubernetes.io/projected/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-kube-api-access-r6ck7\") pod \"redhat-operators-g8xxs\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.672804 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.677410 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:09 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:09 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:09 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.677471 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.780158 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh74p"] Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.850770 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.911721 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vdhp6"] Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.912738 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.925080 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vdhp6"] Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.940221 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-catalog-content\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.940262 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbxkj\" (UniqueName: \"kubernetes.io/projected/49a419d1-7b09-45dd-a1ea-77a97158b16b-kube-api-access-lbxkj\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:09 crc kubenswrapper[4810]: I1008 06:34:09.940295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-utilities\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.041476 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-catalog-content\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.042037 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbxkj\" (UniqueName: \"kubernetes.io/projected/49a419d1-7b09-45dd-a1ea-77a97158b16b-kube-api-access-lbxkj\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.042115 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-utilities\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.042424 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-utilities\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.042483 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-catalog-content\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.053045 4810 generic.go:334] "Generic (PLEG): container finished" podID="432c1370-925d-4d16-9bc0-b7736387e058" containerID="100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151" exitCode=0 Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.053119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtjk2" event={"ID":"432c1370-925d-4d16-9bc0-b7736387e058","Type":"ContainerDied","Data":"100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151"} Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.053151 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtjk2" event={"ID":"432c1370-925d-4d16-9bc0-b7736387e058","Type":"ContainerStarted","Data":"e7537e535a039e14b923a7f1b47e78951de86ecf322f93e2eb410cea19ceb1ca"} Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.057440 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh74p" event={"ID":"87bfa11d-2272-489c-8a91-2b0e22acf8dd","Type":"ContainerStarted","Data":"41354d8defa09c8031e447e0e68be942affc14c97604261a1733045fc6b7b917"} Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.060027 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" event={"ID":"5651941c-e009-48b5-b824-69351bb54c6d","Type":"ContainerDied","Data":"4156052a005e1c72ce5245e9c97043e22fecad18314d28a9c8f310569bfa25ce"} Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.060048 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4156052a005e1c72ce5245e9c97043e22fecad18314d28a9c8f310569bfa25ce" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.060084 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331750-s57sb" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.064823 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbxkj\" (UniqueName: \"kubernetes.io/projected/49a419d1-7b09-45dd-a1ea-77a97158b16b-kube-api-access-lbxkj\") pod \"redhat-operators-vdhp6\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.159203 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8xxs"] Oct 08 06:34:10 crc kubenswrapper[4810]: W1008 06:34:10.188721 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod094d3d4c_b9d5_4d74_a851_7ac814d0f3e6.slice/crio-d42498cfca0e06a790750719bfc9a41e9eee52d3e1f766682b4937f26092b28d WatchSource:0}: Error finding container d42498cfca0e06a790750719bfc9a41e9eee52d3e1f766682b4937f26092b28d: Status 404 returned error can't find the container with id d42498cfca0e06a790750719bfc9a41e9eee52d3e1f766682b4937f26092b28d Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.256504 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.402206 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.448051 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kubelet-dir\") pod \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.448304 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kube-api-access\") pod \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\" (UID: \"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5\") " Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.449156 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5" (UID: "84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.457424 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5" (UID: "84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.549413 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.549449 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.671365 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:10 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:10 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:10 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.671458 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:10 crc kubenswrapper[4810]: I1008 06:34:10.741948 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vdhp6"] Oct 08 06:34:10 crc kubenswrapper[4810]: W1008 06:34:10.812251 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49a419d1_7b09_45dd_a1ea_77a97158b16b.slice/crio-3fdf611a15d4b516c41ff598e9351af833d2fb0d8385108ed0cfec353c557c8c WatchSource:0}: Error finding container 3fdf611a15d4b516c41ff598e9351af833d2fb0d8385108ed0cfec353c557c8c: Status 404 returned error can't find the container with id 3fdf611a15d4b516c41ff598e9351af833d2fb0d8385108ed0cfec353c557c8c Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.068613 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5","Type":"ContainerDied","Data":"717a550925a5bc76fec0f878359b8e0a39c50959a6635b8399577613a05b2645"} Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.068912 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="717a550925a5bc76fec0f878359b8e0a39c50959a6635b8399577613a05b2645" Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.068878 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.070508 4810 generic.go:334] "Generic (PLEG): container finished" podID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerID="a63784b12658cc6fedd2d4b27bf2277ac8eafe6fbfe472af3b7c5eab32b235d9" exitCode=0 Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.070589 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh74p" event={"ID":"87bfa11d-2272-489c-8a91-2b0e22acf8dd","Type":"ContainerDied","Data":"a63784b12658cc6fedd2d4b27bf2277ac8eafe6fbfe472af3b7c5eab32b235d9"} Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.071887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdhp6" event={"ID":"49a419d1-7b09-45dd-a1ea-77a97158b16b","Type":"ContainerStarted","Data":"3fdf611a15d4b516c41ff598e9351af833d2fb0d8385108ed0cfec353c557c8c"} Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.074809 4810 generic.go:334] "Generic (PLEG): container finished" podID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerID="4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c" exitCode=0 Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.074873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8xxs" event={"ID":"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6","Type":"ContainerDied","Data":"4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c"} Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.074908 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8xxs" event={"ID":"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6","Type":"ContainerStarted","Data":"d42498cfca0e06a790750719bfc9a41e9eee52d3e1f766682b4937f26092b28d"} Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.475373 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-752k2" Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.672269 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:11 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:11 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:11 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:11 crc kubenswrapper[4810]: I1008 06:34:11.672347 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.089735 4810 generic.go:334] "Generic (PLEG): container finished" podID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerID="b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83" exitCode=0 Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.089865 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdhp6" event={"ID":"49a419d1-7b09-45dd-a1ea-77a97158b16b","Type":"ContainerDied","Data":"b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83"} Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.137671 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 06:34:12 crc kubenswrapper[4810]: E1008 06:34:12.138049 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5" containerName="pruner" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.138066 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5" containerName="pruner" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.138199 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="84df4a1c-65e8-4eac-ba4e-5e31af4f5ad5" containerName="pruner" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.138811 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.140526 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.141882 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.145605 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.176364 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c1f98771-506a-4c29-babc-82cc73e5db23-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.176441 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1f98771-506a-4c29-babc-82cc73e5db23-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.218455 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.278187 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c1f98771-506a-4c29-babc-82cc73e5db23-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.278317 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1f98771-506a-4c29-babc-82cc73e5db23-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.278385 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c1f98771-506a-4c29-babc-82cc73e5db23-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.305620 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1f98771-506a-4c29-babc-82cc73e5db23-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.467949 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.671858 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:12 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:12 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:12 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:12 crc kubenswrapper[4810]: I1008 06:34:12.671949 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:13 crc kubenswrapper[4810]: I1008 06:34:13.031219 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:34:13 crc kubenswrapper[4810]: I1008 06:34:13.037396 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kt2pp" Oct 08 06:34:13 crc kubenswrapper[4810]: I1008 06:34:13.677354 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:13 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:13 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:13 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:13 crc kubenswrapper[4810]: I1008 06:34:13.677461 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:14 crc kubenswrapper[4810]: I1008 06:34:14.671635 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:14 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:14 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:14 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:14 crc kubenswrapper[4810]: I1008 06:34:14.672281 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:15 crc kubenswrapper[4810]: I1008 06:34:15.671153 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:15 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:15 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:15 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:15 crc kubenswrapper[4810]: I1008 06:34:15.671320 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:16 crc kubenswrapper[4810]: I1008 06:34:16.671022 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:16 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:16 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:16 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:16 crc kubenswrapper[4810]: I1008 06:34:16.671103 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:17 crc kubenswrapper[4810]: I1008 06:34:17.671587 4810 patch_prober.go:28] interesting pod/router-default-5444994796-fwdsd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 06:34:17 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Oct 08 06:34:17 crc kubenswrapper[4810]: [+]process-running ok Oct 08 06:34:17 crc kubenswrapper[4810]: healthz check failed Oct 08 06:34:17 crc kubenswrapper[4810]: I1008 06:34:17.671939 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fwdsd" podUID="646d90cc-1f56-4fd1-bd7f-938021eeb92a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 06:34:17 crc kubenswrapper[4810]: I1008 06:34:17.705048 4810 patch_prober.go:28] interesting pod/console-f9d7485db-fq7wx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 08 06:34:17 crc kubenswrapper[4810]: I1008 06:34:17.705122 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fq7wx" podUID="303e572d-0bdf-4cf2-85bf-976e5ac59130" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 08 06:34:18 crc kubenswrapper[4810]: I1008 06:34:18.040328 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6m85n" Oct 08 06:34:18 crc kubenswrapper[4810]: I1008 06:34:18.673058 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:34:18 crc kubenswrapper[4810]: I1008 06:34:18.677183 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-fwdsd" Oct 08 06:34:19 crc kubenswrapper[4810]: I1008 06:34:19.431694 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:34:19 crc kubenswrapper[4810]: I1008 06:34:19.431797 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:34:24 crc kubenswrapper[4810]: I1008 06:34:24.623213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:34:24 crc kubenswrapper[4810]: I1008 06:34:24.631445 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/669c3fc4-7586-4517-83c9-6316b2671329-metrics-certs\") pod \"network-metrics-daemon-9s2m8\" (UID: \"669c3fc4-7586-4517-83c9-6316b2671329\") " pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:34:24 crc kubenswrapper[4810]: I1008 06:34:24.796415 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9s2m8" Oct 08 06:34:26 crc kubenswrapper[4810]: I1008 06:34:26.536174 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:34:27 crc kubenswrapper[4810]: I1008 06:34:27.707925 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:34:27 crc kubenswrapper[4810]: I1008 06:34:27.713551 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.354708 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.355336 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pccb4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7fw7n_openshift-marketplace(3abe3edb-3d3f-48a0-9968-f47012715666): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.356567 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7fw7n" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.381376 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.381666 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6vm9j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bhgcd_openshift-marketplace(73f9f84f-5f3c-45fa-8b5f-108bd77b2251): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.382876 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-bhgcd" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.996661 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.996798 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b6b8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xtjk2_openshift-marketplace(432c1370-925d-4d16-9bc0-b7736387e058): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 06:34:30 crc kubenswrapper[4810]: E1008 06:34:30.997948 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xtjk2" podUID="432c1370-925d-4d16-9bc0-b7736387e058" Oct 08 06:34:34 crc kubenswrapper[4810]: E1008 06:34:34.068547 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xtjk2" podUID="432c1370-925d-4d16-9bc0-b7736387e058" Oct 08 06:34:34 crc kubenswrapper[4810]: E1008 06:34:34.068675 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7fw7n" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" Oct 08 06:34:34 crc kubenswrapper[4810]: E1008 06:34:34.068703 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bhgcd" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" Oct 08 06:34:35 crc kubenswrapper[4810]: E1008 06:34:35.435790 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 06:34:35 crc kubenswrapper[4810]: E1008 06:34:35.436080 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jgwpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-l98x5_openshift-marketplace(559f47bc-728a-4ef9-bff7-bcda90c4134e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 06:34:35 crc kubenswrapper[4810]: E1008 06:34:35.438140 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-l98x5" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" Oct 08 06:34:35 crc kubenswrapper[4810]: E1008 06:34:35.500763 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 06:34:35 crc kubenswrapper[4810]: E1008 06:34:35.501618 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bzlmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kb4ww_openshift-marketplace(85ab7796-c612-4b27-a367-1d3deca20d7a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 06:34:35 crc kubenswrapper[4810]: E1008 06:34:35.502813 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kb4ww" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" Oct 08 06:34:35 crc kubenswrapper[4810]: I1008 06:34:35.961992 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9s2m8"] Oct 08 06:34:35 crc kubenswrapper[4810]: W1008 06:34:35.967307 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod669c3fc4_7586_4517_83c9_6316b2671329.slice/crio-03b2780ee631a10e0f3cf8e44826fc68480d5f337a83b61581a0fa761f5fa18d WatchSource:0}: Error finding container 03b2780ee631a10e0f3cf8e44826fc68480d5f337a83b61581a0fa761f5fa18d: Status 404 returned error can't find the container with id 03b2780ee631a10e0f3cf8e44826fc68480d5f337a83b61581a0fa761f5fa18d Oct 08 06:34:36 crc kubenswrapper[4810]: I1008 06:34:36.033348 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 06:34:36 crc kubenswrapper[4810]: I1008 06:34:36.257013 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" event={"ID":"669c3fc4-7586-4517-83c9-6316b2671329","Type":"ContainerStarted","Data":"03b2780ee631a10e0f3cf8e44826fc68480d5f337a83b61581a0fa761f5fa18d"} Oct 08 06:34:36 crc kubenswrapper[4810]: I1008 06:34:36.259541 4810 generic.go:334] "Generic (PLEG): container finished" podID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerID="06d2ff3766a3cf77da96388565b28d024a15d67af837928459ca4dacc41a92da" exitCode=0 Oct 08 06:34:36 crc kubenswrapper[4810]: I1008 06:34:36.259637 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh74p" event={"ID":"87bfa11d-2272-489c-8a91-2b0e22acf8dd","Type":"ContainerDied","Data":"06d2ff3766a3cf77da96388565b28d024a15d67af837928459ca4dacc41a92da"} Oct 08 06:34:36 crc kubenswrapper[4810]: I1008 06:34:36.260947 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c1f98771-506a-4c29-babc-82cc73e5db23","Type":"ContainerStarted","Data":"ca15ae5906e70c29e6ed7d72a47598e879a7679a7329265c77b8b3ea53c19800"} Oct 08 06:34:36 crc kubenswrapper[4810]: I1008 06:34:36.275659 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdhp6" event={"ID":"49a419d1-7b09-45dd-a1ea-77a97158b16b","Type":"ContainerStarted","Data":"7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27"} Oct 08 06:34:36 crc kubenswrapper[4810]: I1008 06:34:36.279901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8xxs" event={"ID":"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6","Type":"ContainerStarted","Data":"da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a"} Oct 08 06:34:36 crc kubenswrapper[4810]: E1008 06:34:36.283314 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-l98x5" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" Oct 08 06:34:36 crc kubenswrapper[4810]: E1008 06:34:36.283697 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kb4ww" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.296093 4810 generic.go:334] "Generic (PLEG): container finished" podID="c1f98771-506a-4c29-babc-82cc73e5db23" containerID="9225ee9c42a0fb34a9ccd72111cd6f59707cde8937618a17a68026287b76a6c9" exitCode=0 Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.296201 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c1f98771-506a-4c29-babc-82cc73e5db23","Type":"ContainerDied","Data":"9225ee9c42a0fb34a9ccd72111cd6f59707cde8937618a17a68026287b76a6c9"} Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.302919 4810 generic.go:334] "Generic (PLEG): container finished" podID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerID="7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27" exitCode=0 Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.303033 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdhp6" event={"ID":"49a419d1-7b09-45dd-a1ea-77a97158b16b","Type":"ContainerDied","Data":"7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27"} Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.307240 4810 generic.go:334] "Generic (PLEG): container finished" podID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerID="da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a" exitCode=0 Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.307294 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8xxs" event={"ID":"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6","Type":"ContainerDied","Data":"da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a"} Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.313851 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" event={"ID":"669c3fc4-7586-4517-83c9-6316b2671329","Type":"ContainerStarted","Data":"25dd6e38280e163842b1199e04305362317bc68db356bc932546d136c531e875"} Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.313891 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9s2m8" event={"ID":"669c3fc4-7586-4517-83c9-6316b2671329","Type":"ContainerStarted","Data":"905f4d47f37bfabde5567407a25787b181eb60e013dd5c142d2d56bb4fca7e0b"} Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.327491 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh74p" event={"ID":"87bfa11d-2272-489c-8a91-2b0e22acf8dd","Type":"ContainerStarted","Data":"1eb3cbc30bfef5a1780aa231b1675ea4755ab082430a79e51421aaa532971cd3"} Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.355890 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9s2m8" podStartSLOduration=156.355855023 podStartE2EDuration="2m36.355855023s" podCreationTimestamp="2025-10-08 06:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:34:37.352655217 +0000 UTC m=+179.987094977" watchObservedRunningTime="2025-10-08 06:34:37.355855023 +0000 UTC m=+179.990294803" Oct 08 06:34:37 crc kubenswrapper[4810]: I1008 06:34:37.403548 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xh74p" podStartSLOduration=11.089158171 podStartE2EDuration="29.403511998s" podCreationTimestamp="2025-10-08 06:34:08 +0000 UTC" firstStartedPulling="2025-10-08 06:34:18.396366017 +0000 UTC m=+161.030805757" lastFinishedPulling="2025-10-08 06:34:36.710719804 +0000 UTC m=+179.345159584" observedRunningTime="2025-10-08 06:34:37.396751205 +0000 UTC m=+180.031190995" watchObservedRunningTime="2025-10-08 06:34:37.403511998 +0000 UTC m=+180.037951778" Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.336036 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdhp6" event={"ID":"49a419d1-7b09-45dd-a1ea-77a97158b16b","Type":"ContainerStarted","Data":"abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be"} Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.338451 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8xxs" event={"ID":"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6","Type":"ContainerStarted","Data":"7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77"} Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.363860 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vdhp6" podStartSLOduration=10.006327303 podStartE2EDuration="29.363837588s" podCreationTimestamp="2025-10-08 06:34:09 +0000 UTC" firstStartedPulling="2025-10-08 06:34:18.396412918 +0000 UTC m=+161.030852668" lastFinishedPulling="2025-10-08 06:34:37.753923123 +0000 UTC m=+180.388362953" observedRunningTime="2025-10-08 06:34:38.360436096 +0000 UTC m=+180.994875836" watchObservedRunningTime="2025-10-08 06:34:38.363837588 +0000 UTC m=+180.998277328" Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.389993 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8xxs" podStartSLOduration=2.698911882 podStartE2EDuration="29.389951051s" podCreationTimestamp="2025-10-08 06:34:09 +0000 UTC" firstStartedPulling="2025-10-08 06:34:11.076913858 +0000 UTC m=+153.711353598" lastFinishedPulling="2025-10-08 06:34:37.767953027 +0000 UTC m=+180.402392767" observedRunningTime="2025-10-08 06:34:38.385803934 +0000 UTC m=+181.020243674" watchObservedRunningTime="2025-10-08 06:34:38.389951051 +0000 UTC m=+181.024390791" Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.650483 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.741355 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1f98771-506a-4c29-babc-82cc73e5db23-kube-api-access\") pod \"c1f98771-506a-4c29-babc-82cc73e5db23\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.741492 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c1f98771-506a-4c29-babc-82cc73e5db23-kubelet-dir\") pod \"c1f98771-506a-4c29-babc-82cc73e5db23\" (UID: \"c1f98771-506a-4c29-babc-82cc73e5db23\") " Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.741822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1f98771-506a-4c29-babc-82cc73e5db23-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c1f98771-506a-4c29-babc-82cc73e5db23" (UID: "c1f98771-506a-4c29-babc-82cc73e5db23"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.749256 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1f98771-506a-4c29-babc-82cc73e5db23-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c1f98771-506a-4c29-babc-82cc73e5db23" (UID: "c1f98771-506a-4c29-babc-82cc73e5db23"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.843559 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1f98771-506a-4c29-babc-82cc73e5db23-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:38 crc kubenswrapper[4810]: I1008 06:34:38.843604 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c1f98771-506a-4c29-babc-82cc73e5db23-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.288339 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.288392 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.345525 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c1f98771-506a-4c29-babc-82cc73e5db23","Type":"ContainerDied","Data":"ca15ae5906e70c29e6ed7d72a47598e879a7679a7329265c77b8b3ea53c19800"} Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.345578 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca15ae5906e70c29e6ed7d72a47598e879a7679a7329265c77b8b3ea53c19800" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.345652 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.432477 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.737727 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h2zzg" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.851334 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:39 crc kubenswrapper[4810]: I1008 06:34:39.851391 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:40 crc kubenswrapper[4810]: I1008 06:34:40.257891 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:40 crc kubenswrapper[4810]: I1008 06:34:40.258033 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:40 crc kubenswrapper[4810]: I1008 06:34:40.904840 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g8xxs" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="registry-server" probeResult="failure" output=< Oct 08 06:34:40 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Oct 08 06:34:40 crc kubenswrapper[4810]: > Oct 08 06:34:41 crc kubenswrapper[4810]: I1008 06:34:41.298287 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vdhp6" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="registry-server" probeResult="failure" output=< Oct 08 06:34:41 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Oct 08 06:34:41 crc kubenswrapper[4810]: > Oct 08 06:34:46 crc kubenswrapper[4810]: I1008 06:34:46.423915 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 06:34:49 crc kubenswrapper[4810]: I1008 06:34:49.344311 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:49 crc kubenswrapper[4810]: I1008 06:34:49.431386 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:34:49 crc kubenswrapper[4810]: I1008 06:34:49.431455 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:34:49 crc kubenswrapper[4810]: I1008 06:34:49.899158 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:49 crc kubenswrapper[4810]: I1008 06:34:49.951196 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.299867 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.352993 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.417021 4810 generic.go:334] "Generic (PLEG): container finished" podID="432c1370-925d-4d16-9bc0-b7736387e058" containerID="95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263" exitCode=0 Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.417096 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtjk2" event={"ID":"432c1370-925d-4d16-9bc0-b7736387e058","Type":"ContainerDied","Data":"95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263"} Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.419510 4810 generic.go:334] "Generic (PLEG): container finished" podID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerID="35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387" exitCode=0 Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.419569 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhgcd" event={"ID":"73f9f84f-5f3c-45fa-8b5f-108bd77b2251","Type":"ContainerDied","Data":"35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387"} Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.427877 4810 generic.go:334] "Generic (PLEG): container finished" podID="3abe3edb-3d3f-48a0-9968-f47012715666" containerID="77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6" exitCode=0 Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.427954 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fw7n" event={"ID":"3abe3edb-3d3f-48a0-9968-f47012715666","Type":"ContainerDied","Data":"77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6"} Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.434058 4810 generic.go:334] "Generic (PLEG): container finished" podID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerID="989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc" exitCode=0 Oct 08 06:34:50 crc kubenswrapper[4810]: I1008 06:34:50.434181 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb4ww" event={"ID":"85ab7796-c612-4b27-a367-1d3deca20d7a","Type":"ContainerDied","Data":"989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc"} Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.442184 4810 generic.go:334] "Generic (PLEG): container finished" podID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerID="5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40" exitCode=0 Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.442651 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l98x5" event={"ID":"559f47bc-728a-4ef9-bff7-bcda90c4134e","Type":"ContainerDied","Data":"5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40"} Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.449871 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtjk2" event={"ID":"432c1370-925d-4d16-9bc0-b7736387e058","Type":"ContainerStarted","Data":"9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd"} Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.455142 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhgcd" event={"ID":"73f9f84f-5f3c-45fa-8b5f-108bd77b2251","Type":"ContainerStarted","Data":"f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9"} Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.464454 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb4ww" event={"ID":"85ab7796-c612-4b27-a367-1d3deca20d7a","Type":"ContainerStarted","Data":"49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32"} Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.513788 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xtjk2" podStartSLOduration=2.722519912 podStartE2EDuration="43.513769737s" podCreationTimestamp="2025-10-08 06:34:08 +0000 UTC" firstStartedPulling="2025-10-08 06:34:10.05616879 +0000 UTC m=+152.690608530" lastFinishedPulling="2025-10-08 06:34:50.847418615 +0000 UTC m=+193.481858355" observedRunningTime="2025-10-08 06:34:51.513090848 +0000 UTC m=+194.147530588" watchObservedRunningTime="2025-10-08 06:34:51.513769737 +0000 UTC m=+194.148209477" Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.514064 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kb4ww" podStartSLOduration=2.3115912 podStartE2EDuration="45.514060245s" podCreationTimestamp="2025-10-08 06:34:06 +0000 UTC" firstStartedPulling="2025-10-08 06:34:08.010061053 +0000 UTC m=+150.644500793" lastFinishedPulling="2025-10-08 06:34:51.212530098 +0000 UTC m=+193.846969838" observedRunningTime="2025-10-08 06:34:51.496225379 +0000 UTC m=+194.130665129" watchObservedRunningTime="2025-10-08 06:34:51.514060245 +0000 UTC m=+194.148499985" Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.539798 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bhgcd" podStartSLOduration=2.565014567 podStartE2EDuration="45.539780879s" podCreationTimestamp="2025-10-08 06:34:06 +0000 UTC" firstStartedPulling="2025-10-08 06:34:08.022881637 +0000 UTC m=+150.657321387" lastFinishedPulling="2025-10-08 06:34:50.997647969 +0000 UTC m=+193.632087699" observedRunningTime="2025-10-08 06:34:51.539494601 +0000 UTC m=+194.173934341" watchObservedRunningTime="2025-10-08 06:34:51.539780879 +0000 UTC m=+194.174220619" Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.728195 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh74p"] Oct 08 06:34:51 crc kubenswrapper[4810]: I1008 06:34:51.728516 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xh74p" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="registry-server" containerID="cri-o://1eb3cbc30bfef5a1780aa231b1675ea4755ab082430a79e51421aaa532971cd3" gracePeriod=2 Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.473386 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fw7n" event={"ID":"3abe3edb-3d3f-48a0-9968-f47012715666","Type":"ContainerStarted","Data":"aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe"} Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.481851 4810 generic.go:334] "Generic (PLEG): container finished" podID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerID="1eb3cbc30bfef5a1780aa231b1675ea4755ab082430a79e51421aaa532971cd3" exitCode=0 Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.481905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh74p" event={"ID":"87bfa11d-2272-489c-8a91-2b0e22acf8dd","Type":"ContainerDied","Data":"1eb3cbc30bfef5a1780aa231b1675ea4755ab082430a79e51421aaa532971cd3"} Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.500124 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7fw7n" podStartSLOduration=3.197786189 podStartE2EDuration="46.500104068s" podCreationTimestamp="2025-10-08 06:34:06 +0000 UTC" firstStartedPulling="2025-10-08 06:34:08.000759795 +0000 UTC m=+150.635199555" lastFinishedPulling="2025-10-08 06:34:51.303077694 +0000 UTC m=+193.937517434" observedRunningTime="2025-10-08 06:34:52.497912576 +0000 UTC m=+195.132352316" watchObservedRunningTime="2025-10-08 06:34:52.500104068 +0000 UTC m=+195.134543808" Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.901340 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.988118 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-catalog-content\") pod \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.988269 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42pz5\" (UniqueName: \"kubernetes.io/projected/87bfa11d-2272-489c-8a91-2b0e22acf8dd-kube-api-access-42pz5\") pod \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.988485 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-utilities\") pod \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\" (UID: \"87bfa11d-2272-489c-8a91-2b0e22acf8dd\") " Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.989253 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-utilities" (OuterVolumeSpecName: "utilities") pod "87bfa11d-2272-489c-8a91-2b0e22acf8dd" (UID: "87bfa11d-2272-489c-8a91-2b0e22acf8dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:34:52 crc kubenswrapper[4810]: I1008 06:34:52.996155 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87bfa11d-2272-489c-8a91-2b0e22acf8dd-kube-api-access-42pz5" (OuterVolumeSpecName: "kube-api-access-42pz5") pod "87bfa11d-2272-489c-8a91-2b0e22acf8dd" (UID: "87bfa11d-2272-489c-8a91-2b0e22acf8dd"). InnerVolumeSpecName "kube-api-access-42pz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.002916 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87bfa11d-2272-489c-8a91-2b0e22acf8dd" (UID: "87bfa11d-2272-489c-8a91-2b0e22acf8dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.090267 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.090327 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42pz5\" (UniqueName: \"kubernetes.io/projected/87bfa11d-2272-489c-8a91-2b0e22acf8dd-kube-api-access-42pz5\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.090350 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87bfa11d-2272-489c-8a91-2b0e22acf8dd-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.129664 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vdhp6"] Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.130365 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vdhp6" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="registry-server" containerID="cri-o://abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be" gracePeriod=2 Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.496087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xh74p" event={"ID":"87bfa11d-2272-489c-8a91-2b0e22acf8dd","Type":"ContainerDied","Data":"41354d8defa09c8031e447e0e68be942affc14c97604261a1733045fc6b7b917"} Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.496138 4810 scope.go:117] "RemoveContainer" containerID="1eb3cbc30bfef5a1780aa231b1675ea4755ab082430a79e51421aaa532971cd3" Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.496270 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xh74p" Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.527175 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh74p"] Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.532491 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xh74p"] Oct 08 06:34:53 crc kubenswrapper[4810]: I1008 06:34:53.882617 4810 scope.go:117] "RemoveContainer" containerID="06d2ff3766a3cf77da96388565b28d024a15d67af837928459ca4dacc41a92da" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.015916 4810 scope.go:117] "RemoveContainer" containerID="a63784b12658cc6fedd2d4b27bf2277ac8eafe6fbfe472af3b7c5eab32b235d9" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.085998 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" path="/var/lib/kubelet/pods/87bfa11d-2272-489c-8a91-2b0e22acf8dd/volumes" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.336445 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.412855 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-catalog-content\") pod \"49a419d1-7b09-45dd-a1ea-77a97158b16b\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.412953 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-utilities\") pod \"49a419d1-7b09-45dd-a1ea-77a97158b16b\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.413014 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbxkj\" (UniqueName: \"kubernetes.io/projected/49a419d1-7b09-45dd-a1ea-77a97158b16b-kube-api-access-lbxkj\") pod \"49a419d1-7b09-45dd-a1ea-77a97158b16b\" (UID: \"49a419d1-7b09-45dd-a1ea-77a97158b16b\") " Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.414583 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-utilities" (OuterVolumeSpecName: "utilities") pod "49a419d1-7b09-45dd-a1ea-77a97158b16b" (UID: "49a419d1-7b09-45dd-a1ea-77a97158b16b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.420609 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49a419d1-7b09-45dd-a1ea-77a97158b16b-kube-api-access-lbxkj" (OuterVolumeSpecName: "kube-api-access-lbxkj") pod "49a419d1-7b09-45dd-a1ea-77a97158b16b" (UID: "49a419d1-7b09-45dd-a1ea-77a97158b16b"). InnerVolumeSpecName "kube-api-access-lbxkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.505328 4810 generic.go:334] "Generic (PLEG): container finished" podID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerID="abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be" exitCode=0 Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.505389 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdhp6" event={"ID":"49a419d1-7b09-45dd-a1ea-77a97158b16b","Type":"ContainerDied","Data":"abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be"} Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.505441 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdhp6" event={"ID":"49a419d1-7b09-45dd-a1ea-77a97158b16b","Type":"ContainerDied","Data":"3fdf611a15d4b516c41ff598e9351af833d2fb0d8385108ed0cfec353c557c8c"} Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.505468 4810 scope.go:117] "RemoveContainer" containerID="abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.508136 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdhp6" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.514855 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.515192 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbxkj\" (UniqueName: \"kubernetes.io/projected/49a419d1-7b09-45dd-a1ea-77a97158b16b-kube-api-access-lbxkj\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.524414 4810 scope.go:117] "RemoveContainer" containerID="7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.531248 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49a419d1-7b09-45dd-a1ea-77a97158b16b" (UID: "49a419d1-7b09-45dd-a1ea-77a97158b16b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.553042 4810 scope.go:117] "RemoveContainer" containerID="b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.572622 4810 scope.go:117] "RemoveContainer" containerID="abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be" Oct 08 06:34:54 crc kubenswrapper[4810]: E1008 06:34:54.573231 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be\": container with ID starting with abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be not found: ID does not exist" containerID="abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.573293 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be"} err="failed to get container status \"abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be\": rpc error: code = NotFound desc = could not find container \"abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be\": container with ID starting with abd6038e91cd49f1cdf8cc97872c531b53137f4dd9ac2428597ff8210ffdb1be not found: ID does not exist" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.573365 4810 scope.go:117] "RemoveContainer" containerID="7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27" Oct 08 06:34:54 crc kubenswrapper[4810]: E1008 06:34:54.573957 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27\": container with ID starting with 7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27 not found: ID does not exist" containerID="7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.574073 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27"} err="failed to get container status \"7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27\": rpc error: code = NotFound desc = could not find container \"7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27\": container with ID starting with 7fda1c8a35dd64ac8a3be7916c97b02f5ccb0d48d1ccd7bf215f73493d087e27 not found: ID does not exist" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.574135 4810 scope.go:117] "RemoveContainer" containerID="b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83" Oct 08 06:34:54 crc kubenswrapper[4810]: E1008 06:34:54.574610 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83\": container with ID starting with b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83 not found: ID does not exist" containerID="b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.574650 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83"} err="failed to get container status \"b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83\": rpc error: code = NotFound desc = could not find container \"b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83\": container with ID starting with b3c1fe345fa520fda1d6448519da57a71c3de7dcea7f908a9f97a471091a3c83 not found: ID does not exist" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.617106 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49a419d1-7b09-45dd-a1ea-77a97158b16b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.843100 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vdhp6"] Oct 08 06:34:54 crc kubenswrapper[4810]: I1008 06:34:54.846503 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vdhp6"] Oct 08 06:34:55 crc kubenswrapper[4810]: I1008 06:34:55.513666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l98x5" event={"ID":"559f47bc-728a-4ef9-bff7-bcda90c4134e","Type":"ContainerStarted","Data":"e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b"} Oct 08 06:34:55 crc kubenswrapper[4810]: I1008 06:34:55.531994 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l98x5" podStartSLOduration=3.518365182 podStartE2EDuration="49.531978811s" podCreationTimestamp="2025-10-08 06:34:06 +0000 UTC" firstStartedPulling="2025-10-08 06:34:08.002511754 +0000 UTC m=+150.636951484" lastFinishedPulling="2025-10-08 06:34:54.016125373 +0000 UTC m=+196.650565113" observedRunningTime="2025-10-08 06:34:55.530323523 +0000 UTC m=+198.164763263" watchObservedRunningTime="2025-10-08 06:34:55.531978811 +0000 UTC m=+198.166418551" Oct 08 06:34:56 crc kubenswrapper[4810]: I1008 06:34:56.081347 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" path="/var/lib/kubelet/pods/49a419d1-7b09-45dd-a1ea-77a97158b16b/volumes" Oct 08 06:34:56 crc kubenswrapper[4810]: I1008 06:34:56.658410 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:56 crc kubenswrapper[4810]: I1008 06:34:56.658478 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:56 crc kubenswrapper[4810]: I1008 06:34:56.704716 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:56 crc kubenswrapper[4810]: I1008 06:34:56.849227 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:56 crc kubenswrapper[4810]: I1008 06:34:56.849272 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:56 crc kubenswrapper[4810]: I1008 06:34:56.889433 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.063250 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.063298 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.288913 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.289003 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.353857 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.587645 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.595623 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:34:57 crc kubenswrapper[4810]: I1008 06:34:57.597023 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:34:58 crc kubenswrapper[4810]: I1008 06:34:58.110373 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-l98x5" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="registry-server" probeResult="failure" output=< Oct 08 06:34:58 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Oct 08 06:34:58 crc kubenswrapper[4810]: > Oct 08 06:34:58 crc kubenswrapper[4810]: I1008 06:34:58.879867 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:58 crc kubenswrapper[4810]: I1008 06:34:58.880190 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:58 crc kubenswrapper[4810]: I1008 06:34:58.943527 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:59 crc kubenswrapper[4810]: I1008 06:34:59.527260 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bhgcd"] Oct 08 06:34:59 crc kubenswrapper[4810]: I1008 06:34:59.541423 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bhgcd" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="registry-server" containerID="cri-o://f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9" gracePeriod=2 Oct 08 06:34:59 crc kubenswrapper[4810]: I1008 06:34:59.597952 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:34:59 crc kubenswrapper[4810]: I1008 06:34:59.973154 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.004484 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-utilities\") pod \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.004567 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vm9j\" (UniqueName: \"kubernetes.io/projected/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-kube-api-access-6vm9j\") pod \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.004622 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-catalog-content\") pod \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\" (UID: \"73f9f84f-5f3c-45fa-8b5f-108bd77b2251\") " Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.006687 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-utilities" (OuterVolumeSpecName: "utilities") pod "73f9f84f-5f3c-45fa-8b5f-108bd77b2251" (UID: "73f9f84f-5f3c-45fa-8b5f-108bd77b2251"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.014472 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-kube-api-access-6vm9j" (OuterVolumeSpecName: "kube-api-access-6vm9j") pod "73f9f84f-5f3c-45fa-8b5f-108bd77b2251" (UID: "73f9f84f-5f3c-45fa-8b5f-108bd77b2251"). InnerVolumeSpecName "kube-api-access-6vm9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.073882 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73f9f84f-5f3c-45fa-8b5f-108bd77b2251" (UID: "73f9f84f-5f3c-45fa-8b5f-108bd77b2251"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.106762 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.106797 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.106823 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vm9j\" (UniqueName: \"kubernetes.io/projected/73f9f84f-5f3c-45fa-8b5f-108bd77b2251-kube-api-access-6vm9j\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.170504 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sxjjh"] Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.550828 4810 generic.go:334] "Generic (PLEG): container finished" podID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerID="f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9" exitCode=0 Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.550913 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhgcd" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.550925 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhgcd" event={"ID":"73f9f84f-5f3c-45fa-8b5f-108bd77b2251","Type":"ContainerDied","Data":"f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9"} Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.551046 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhgcd" event={"ID":"73f9f84f-5f3c-45fa-8b5f-108bd77b2251","Type":"ContainerDied","Data":"608a272c8f1572a5bbeff3b2e75fdbc0b9de2dab32cf28dba12f5a3b2e3fde79"} Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.551133 4810 scope.go:117] "RemoveContainer" containerID="f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.578705 4810 scope.go:117] "RemoveContainer" containerID="35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.585236 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bhgcd"] Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.590832 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bhgcd"] Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.600367 4810 scope.go:117] "RemoveContainer" containerID="eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.620321 4810 scope.go:117] "RemoveContainer" containerID="f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9" Oct 08 06:35:00 crc kubenswrapper[4810]: E1008 06:35:00.621151 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9\": container with ID starting with f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9 not found: ID does not exist" containerID="f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.621209 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9"} err="failed to get container status \"f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9\": rpc error: code = NotFound desc = could not find container \"f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9\": container with ID starting with f03688f0dc79a054b5b5dea0ee067627af481b7f8f0833111fecf40c18548da9 not found: ID does not exist" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.621255 4810 scope.go:117] "RemoveContainer" containerID="35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387" Oct 08 06:35:00 crc kubenswrapper[4810]: E1008 06:35:00.622167 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387\": container with ID starting with 35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387 not found: ID does not exist" containerID="35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.622247 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387"} err="failed to get container status \"35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387\": rpc error: code = NotFound desc = could not find container \"35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387\": container with ID starting with 35311089e9b47323fc5f7efc8dae2424dddeb259aa49d5bc9ba7a9966e369387 not found: ID does not exist" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.622346 4810 scope.go:117] "RemoveContainer" containerID="eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5" Oct 08 06:35:00 crc kubenswrapper[4810]: E1008 06:35:00.622903 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5\": container with ID starting with eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5 not found: ID does not exist" containerID="eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5" Oct 08 06:35:00 crc kubenswrapper[4810]: I1008 06:35:00.622981 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5"} err="failed to get container status \"eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5\": rpc error: code = NotFound desc = could not find container \"eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5\": container with ID starting with eb28793af7bd390a8916ea235904bb9ca7c9804022177d06289fbfa5bf934db5 not found: ID does not exist" Oct 08 06:35:02 crc kubenswrapper[4810]: I1008 06:35:02.080248 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" path="/var/lib/kubelet/pods/73f9f84f-5f3c-45fa-8b5f-108bd77b2251/volumes" Oct 08 06:35:07 crc kubenswrapper[4810]: I1008 06:35:07.097254 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:35:07 crc kubenswrapper[4810]: I1008 06:35:07.137930 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:35:07 crc kubenswrapper[4810]: I1008 06:35:07.933849 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l98x5"] Oct 08 06:35:08 crc kubenswrapper[4810]: I1008 06:35:08.597786 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l98x5" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="registry-server" containerID="cri-o://e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b" gracePeriod=2 Oct 08 06:35:08 crc kubenswrapper[4810]: I1008 06:35:08.993741 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.031574 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-catalog-content\") pod \"559f47bc-728a-4ef9-bff7-bcda90c4134e\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.031653 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgwpf\" (UniqueName: \"kubernetes.io/projected/559f47bc-728a-4ef9-bff7-bcda90c4134e-kube-api-access-jgwpf\") pod \"559f47bc-728a-4ef9-bff7-bcda90c4134e\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.031684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-utilities\") pod \"559f47bc-728a-4ef9-bff7-bcda90c4134e\" (UID: \"559f47bc-728a-4ef9-bff7-bcda90c4134e\") " Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.032615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-utilities" (OuterVolumeSpecName: "utilities") pod "559f47bc-728a-4ef9-bff7-bcda90c4134e" (UID: "559f47bc-728a-4ef9-bff7-bcda90c4134e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.039014 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/559f47bc-728a-4ef9-bff7-bcda90c4134e-kube-api-access-jgwpf" (OuterVolumeSpecName: "kube-api-access-jgwpf") pod "559f47bc-728a-4ef9-bff7-bcda90c4134e" (UID: "559f47bc-728a-4ef9-bff7-bcda90c4134e"). InnerVolumeSpecName "kube-api-access-jgwpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.081117 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "559f47bc-728a-4ef9-bff7-bcda90c4134e" (UID: "559f47bc-728a-4ef9-bff7-bcda90c4134e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.133248 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgwpf\" (UniqueName: \"kubernetes.io/projected/559f47bc-728a-4ef9-bff7-bcda90c4134e-kube-api-access-jgwpf\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.133289 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.133299 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559f47bc-728a-4ef9-bff7-bcda90c4134e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.606028 4810 generic.go:334] "Generic (PLEG): container finished" podID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerID="e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b" exitCode=0 Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.606081 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l98x5" event={"ID":"559f47bc-728a-4ef9-bff7-bcda90c4134e","Type":"ContainerDied","Data":"e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b"} Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.606121 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l98x5" event={"ID":"559f47bc-728a-4ef9-bff7-bcda90c4134e","Type":"ContainerDied","Data":"0053b534a76f7d8d5f287cbe7c7df7965d77a0260637956ec06d975f4b3aafa8"} Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.606143 4810 scope.go:117] "RemoveContainer" containerID="e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.606229 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l98x5" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.676319 4810 scope.go:117] "RemoveContainer" containerID="5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.676709 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l98x5"] Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.680682 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l98x5"] Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.693232 4810 scope.go:117] "RemoveContainer" containerID="151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.707118 4810 scope.go:117] "RemoveContainer" containerID="e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b" Oct 08 06:35:09 crc kubenswrapper[4810]: E1008 06:35:09.707531 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b\": container with ID starting with e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b not found: ID does not exist" containerID="e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.707575 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b"} err="failed to get container status \"e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b\": rpc error: code = NotFound desc = could not find container \"e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b\": container with ID starting with e6a3d9e354e8d9e3bcc545a2a914a35fd99f4b4d04e0c24c0b478db97a26293b not found: ID does not exist" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.707605 4810 scope.go:117] "RemoveContainer" containerID="5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40" Oct 08 06:35:09 crc kubenswrapper[4810]: E1008 06:35:09.707896 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40\": container with ID starting with 5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40 not found: ID does not exist" containerID="5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.707929 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40"} err="failed to get container status \"5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40\": rpc error: code = NotFound desc = could not find container \"5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40\": container with ID starting with 5cd3041b6a418ea5b9900719d5ca9a18ee9c739c2fdbf15f77442c78d0511d40 not found: ID does not exist" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.707949 4810 scope.go:117] "RemoveContainer" containerID="151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c" Oct 08 06:35:09 crc kubenswrapper[4810]: E1008 06:35:09.708333 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c\": container with ID starting with 151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c not found: ID does not exist" containerID="151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c" Oct 08 06:35:09 crc kubenswrapper[4810]: I1008 06:35:09.708398 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c"} err="failed to get container status \"151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c\": rpc error: code = NotFound desc = could not find container \"151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c\": container with ID starting with 151eed8d8cb1754bcd0d5431b027545ba02959a3bc8835f6530a31b88f4c547c not found: ID does not exist" Oct 08 06:35:10 crc kubenswrapper[4810]: I1008 06:35:10.080255 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" path="/var/lib/kubelet/pods/559f47bc-728a-4ef9-bff7-bcda90c4134e/volumes" Oct 08 06:35:19 crc kubenswrapper[4810]: I1008 06:35:19.432059 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:35:19 crc kubenswrapper[4810]: I1008 06:35:19.432466 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:35:19 crc kubenswrapper[4810]: I1008 06:35:19.432543 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:35:19 crc kubenswrapper[4810]: I1008 06:35:19.433471 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:35:19 crc kubenswrapper[4810]: I1008 06:35:19.433577 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea" gracePeriod=600 Oct 08 06:35:19 crc kubenswrapper[4810]: I1008 06:35:19.670774 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea" exitCode=0 Oct 08 06:35:19 crc kubenswrapper[4810]: I1008 06:35:19.670857 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea"} Oct 08 06:35:20 crc kubenswrapper[4810]: I1008 06:35:20.680462 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"fba6723b31f4f6438b0e473e38431e96175c272a5ec8aa181205f38fc5e27943"} Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.201311 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" podUID="c578e5e6-b81a-490e-ad3c-9bfe81172b1b" containerName="oauth-openshift" containerID="cri-o://64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9" gracePeriod=15 Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.620169 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665208 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-92v9d"] Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665581 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c578e5e6-b81a-490e-ad3c-9bfe81172b1b" containerName="oauth-openshift" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665597 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c578e5e6-b81a-490e-ad3c-9bfe81172b1b" containerName="oauth-openshift" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665616 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665624 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665636 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665644 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665655 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665662 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665671 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665679 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665690 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665698 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665710 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665718 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665729 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665738 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="extract-content" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665751 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665761 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665772 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1f98771-506a-4c29-babc-82cc73e5db23" containerName="pruner" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665780 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1f98771-506a-4c29-babc-82cc73e5db23" containerName="pruner" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665791 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665800 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665814 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665822 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665833 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665841 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="extract-utilities" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.665853 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.665861 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.666007 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1f98771-506a-4c29-babc-82cc73e5db23" containerName="pruner" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.666029 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="87bfa11d-2272-489c-8a91-2b0e22acf8dd" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.666041 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a419d1-7b09-45dd-a1ea-77a97158b16b" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.666052 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="559f47bc-728a-4ef9-bff7-bcda90c4134e" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.666068 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f9f84f-5f3c-45fa-8b5f-108bd77b2251" containerName="registry-server" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.666079 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c578e5e6-b81a-490e-ad3c-9bfe81172b1b" containerName="oauth-openshift" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.666593 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.668316 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-92v9d"] Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.725380 4810 generic.go:334] "Generic (PLEG): container finished" podID="c578e5e6-b81a-490e-ad3c-9bfe81172b1b" containerID="64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9" exitCode=0 Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.725426 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" event={"ID":"c578e5e6-b81a-490e-ad3c-9bfe81172b1b","Type":"ContainerDied","Data":"64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9"} Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.725455 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" event={"ID":"c578e5e6-b81a-490e-ad3c-9bfe81172b1b","Type":"ContainerDied","Data":"81ca4db33a9b5c7e3ebbceac635530bd9d66c3d2a094e27b3e1bffb8bd5625e1"} Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.725473 4810 scope.go:117] "RemoveContainer" containerID="64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.725510 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-sxjjh" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.746463 4810 scope.go:117] "RemoveContainer" containerID="64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9" Oct 08 06:35:25 crc kubenswrapper[4810]: E1008 06:35:25.747076 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9\": container with ID starting with 64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9 not found: ID does not exist" containerID="64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.747205 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9"} err="failed to get container status \"64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9\": rpc error: code = NotFound desc = could not find container \"64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9\": container with ID starting with 64170d95987dd5aade4846ec16daf0647e191cb80877104e3bd6eafaf054acd9 not found: ID does not exist" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.779876 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-ocp-branding-template\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-policies\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780281 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-trusted-ca-bundle\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780366 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-error\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780447 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-cliconfig\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780514 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-provider-selection\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-service-ca\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780696 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-router-certs\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780769 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmwbb\" (UniqueName: \"kubernetes.io/projected/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-kube-api-access-fmwbb\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780841 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-idp-0-file-data\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.780929 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-serving-cert\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-dir\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781158 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-login\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781297 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-session\") pod \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\" (UID: \"c578e5e6-b81a-490e-ad3c-9bfe81172b1b\") " Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75m9x\" (UniqueName: \"kubernetes.io/projected/705a681c-0bcc-4907-bf3e-8108755eee7e-kube-api-access-75m9x\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781657 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/705a681c-0bcc-4907-bf3e-8108755eee7e-audit-dir\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781321 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781328 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781735 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.781904 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782107 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782187 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-audit-policies\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782271 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782351 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782456 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782720 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782828 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.782938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.783082 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.783163 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.783269 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.783318 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.783909 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.788565 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.789435 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-kube-api-access-fmwbb" (OuterVolumeSpecName: "kube-api-access-fmwbb") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "kube-api-access-fmwbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.789813 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.790379 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.790616 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.793317 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.800248 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.804648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.805077 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c578e5e6-b81a-490e-ad3c-9bfe81172b1b" (UID: "c578e5e6-b81a-490e-ad3c-9bfe81172b1b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884466 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75m9x\" (UniqueName: \"kubernetes.io/projected/705a681c-0bcc-4907-bf3e-8108755eee7e-kube-api-access-75m9x\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884749 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/705a681c-0bcc-4907-bf3e-8108755eee7e-audit-dir\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884830 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884913 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.884991 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885030 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-audit-policies\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885092 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885136 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885227 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885330 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885358 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885381 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885403 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885427 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885450 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmwbb\" (UniqueName: \"kubernetes.io/projected/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-kube-api-access-fmwbb\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885472 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885498 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885521 4810 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885544 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885566 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.885588 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c578e5e6-b81a-490e-ad3c-9bfe81172b1b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.886040 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/705a681c-0bcc-4907-bf3e-8108755eee7e-audit-dir\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.886725 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-audit-policies\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.886998 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.887301 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.887761 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.890556 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.890665 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.890855 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.890900 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.891374 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.892045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.893434 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.893844 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/705a681c-0bcc-4907-bf3e-8108755eee7e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.910461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75m9x\" (UniqueName: \"kubernetes.io/projected/705a681c-0bcc-4907-bf3e-8108755eee7e-kube-api-access-75m9x\") pod \"oauth-openshift-79656f7ff7-92v9d\" (UID: \"705a681c-0bcc-4907-bf3e-8108755eee7e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:25 crc kubenswrapper[4810]: I1008 06:35:25.984863 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.058752 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sxjjh"] Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.062083 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-sxjjh"] Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.089871 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c578e5e6-b81a-490e-ad3c-9bfe81172b1b" path="/var/lib/kubelet/pods/c578e5e6-b81a-490e-ad3c-9bfe81172b1b/volumes" Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.234957 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-92v9d"] Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.732455 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" event={"ID":"705a681c-0bcc-4907-bf3e-8108755eee7e","Type":"ContainerStarted","Data":"bf80ffcdd125730603a6428967ad21e3012e3fdce65e9e243e575989dcd549c2"} Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.732767 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" event={"ID":"705a681c-0bcc-4907-bf3e-8108755eee7e","Type":"ContainerStarted","Data":"fa555819f3d8aa16b4569e76101bae1b4e26fb2267acbe2fc5665964ce4d8b06"} Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.732787 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.761292 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" podStartSLOduration=26.761247606 podStartE2EDuration="26.761247606s" podCreationTimestamp="2025-10-08 06:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:35:26.752531033 +0000 UTC m=+229.386970803" watchObservedRunningTime="2025-10-08 06:35:26.761247606 +0000 UTC m=+229.395687376" Oct 08 06:35:26 crc kubenswrapper[4810]: I1008 06:35:26.976590 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-79656f7ff7-92v9d" Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.806534 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7fw7n"] Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.808866 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7fw7n" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="registry-server" containerID="cri-o://aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe" gracePeriod=30 Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.814703 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kb4ww"] Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.814899 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kb4ww" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="registry-server" containerID="cri-o://49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32" gracePeriod=30 Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.827004 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7q29k"] Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.827276 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" podUID="74c3e024-0dfe-4f9c-930e-70896c3b8110" containerName="marketplace-operator" containerID="cri-o://a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d" gracePeriod=30 Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.833814 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xtjk2"] Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.836245 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xtjk2" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="registry-server" containerID="cri-o://9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd" gracePeriod=30 Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.847419 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8xxs"] Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.847674 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8xxs" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="registry-server" containerID="cri-o://7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" gracePeriod=30 Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.855500 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k4b69"] Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.857045 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:39 crc kubenswrapper[4810]: E1008 06:35:39.858202 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 06:35:39 crc kubenswrapper[4810]: E1008 06:35:39.862434 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 06:35:39 crc kubenswrapper[4810]: E1008 06:35:39.864096 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 06:35:39 crc kubenswrapper[4810]: E1008 06:35:39.864163 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-operators-g8xxs" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="registry-server" Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.871548 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k4b69"] Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.980035 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2f9ae171-32b7-4ca1-8d91-399067439899-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.980469 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f9ae171-32b7-4ca1-8d91-399067439899-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:39 crc kubenswrapper[4810]: I1008 06:35:39.980527 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9qnc\" (UniqueName: \"kubernetes.io/projected/2f9ae171-32b7-4ca1-8d91-399067439899-kube-api-access-s9qnc\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.082813 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9qnc\" (UniqueName: \"kubernetes.io/projected/2f9ae171-32b7-4ca1-8d91-399067439899-kube-api-access-s9qnc\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.082902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2f9ae171-32b7-4ca1-8d91-399067439899-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.082958 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f9ae171-32b7-4ca1-8d91-399067439899-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.084726 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f9ae171-32b7-4ca1-8d91-399067439899-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.092083 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2f9ae171-32b7-4ca1-8d91-399067439899-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.111313 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9qnc\" (UniqueName: \"kubernetes.io/projected/2f9ae171-32b7-4ca1-8d91-399067439899-kube-api-access-s9qnc\") pod \"marketplace-operator-79b997595-k4b69\" (UID: \"2f9ae171-32b7-4ca1-8d91-399067439899\") " pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.212335 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.283105 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.372435 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.386477 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pccb4\" (UniqueName: \"kubernetes.io/projected/3abe3edb-3d3f-48a0-9968-f47012715666-kube-api-access-pccb4\") pod \"3abe3edb-3d3f-48a0-9968-f47012715666\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.386582 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-utilities\") pod \"3abe3edb-3d3f-48a0-9968-f47012715666\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.386690 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-catalog-content\") pod \"3abe3edb-3d3f-48a0-9968-f47012715666\" (UID: \"3abe3edb-3d3f-48a0-9968-f47012715666\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.391309 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-utilities" (OuterVolumeSpecName: "utilities") pod "3abe3edb-3d3f-48a0-9968-f47012715666" (UID: "3abe3edb-3d3f-48a0-9968-f47012715666"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.396608 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3abe3edb-3d3f-48a0-9968-f47012715666-kube-api-access-pccb4" (OuterVolumeSpecName: "kube-api-access-pccb4") pod "3abe3edb-3d3f-48a0-9968-f47012715666" (UID: "3abe3edb-3d3f-48a0-9968-f47012715666"). InnerVolumeSpecName "kube-api-access-pccb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.401896 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.402346 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.427292 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.482238 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3abe3edb-3d3f-48a0-9968-f47012715666" (UID: "3abe3edb-3d3f-48a0-9968-f47012715666"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.487286 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-utilities\") pod \"85ab7796-c612-4b27-a367-1d3deca20d7a\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.487403 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-catalog-content\") pod \"85ab7796-c612-4b27-a367-1d3deca20d7a\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.487428 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzlmw\" (UniqueName: \"kubernetes.io/projected/85ab7796-c612-4b27-a367-1d3deca20d7a-kube-api-access-bzlmw\") pod \"85ab7796-c612-4b27-a367-1d3deca20d7a\" (UID: \"85ab7796-c612-4b27-a367-1d3deca20d7a\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.487634 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.487644 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pccb4\" (UniqueName: \"kubernetes.io/projected/3abe3edb-3d3f-48a0-9968-f47012715666-kube-api-access-pccb4\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.487653 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3abe3edb-3d3f-48a0-9968-f47012715666-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.489549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-utilities" (OuterVolumeSpecName: "utilities") pod "85ab7796-c612-4b27-a367-1d3deca20d7a" (UID: "85ab7796-c612-4b27-a367-1d3deca20d7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.491380 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85ab7796-c612-4b27-a367-1d3deca20d7a-kube-api-access-bzlmw" (OuterVolumeSpecName: "kube-api-access-bzlmw") pod "85ab7796-c612-4b27-a367-1d3deca20d7a" (UID: "85ab7796-c612-4b27-a367-1d3deca20d7a"). InnerVolumeSpecName "kube-api-access-bzlmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.511704 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k4b69"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.566037 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85ab7796-c612-4b27-a367-1d3deca20d7a" (UID: "85ab7796-c612-4b27-a367-1d3deca20d7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.588841 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6ck7\" (UniqueName: \"kubernetes.io/projected/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-kube-api-access-r6ck7\") pod \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.588889 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-operator-metrics\") pod \"74c3e024-0dfe-4f9c-930e-70896c3b8110\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.588919 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-trusted-ca\") pod \"74c3e024-0dfe-4f9c-930e-70896c3b8110\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.588995 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-utilities\") pod \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589035 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-catalog-content\") pod \"432c1370-925d-4d16-9bc0-b7736387e058\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589069 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-catalog-content\") pod \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\" (UID: \"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589091 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-utilities\") pod \"432c1370-925d-4d16-9bc0-b7736387e058\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589108 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6b8b\" (UniqueName: \"kubernetes.io/projected/432c1370-925d-4d16-9bc0-b7736387e058-kube-api-access-b6b8b\") pod \"432c1370-925d-4d16-9bc0-b7736387e058\" (UID: \"432c1370-925d-4d16-9bc0-b7736387e058\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589144 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjhww\" (UniqueName: \"kubernetes.io/projected/74c3e024-0dfe-4f9c-930e-70896c3b8110-kube-api-access-hjhww\") pod \"74c3e024-0dfe-4f9c-930e-70896c3b8110\" (UID: \"74c3e024-0dfe-4f9c-930e-70896c3b8110\") " Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589360 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589377 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzlmw\" (UniqueName: \"kubernetes.io/projected/85ab7796-c612-4b27-a367-1d3deca20d7a-kube-api-access-bzlmw\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.589387 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85ab7796-c612-4b27-a367-1d3deca20d7a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.591019 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-utilities" (OuterVolumeSpecName: "utilities") pod "432c1370-925d-4d16-9bc0-b7736387e058" (UID: "432c1370-925d-4d16-9bc0-b7736387e058"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.591671 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "74c3e024-0dfe-4f9c-930e-70896c3b8110" (UID: "74c3e024-0dfe-4f9c-930e-70896c3b8110"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.593990 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-utilities" (OuterVolumeSpecName: "utilities") pod "094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" (UID: "094d3d4c-b9d5-4d74-a851-7ac814d0f3e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.596103 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/432c1370-925d-4d16-9bc0-b7736387e058-kube-api-access-b6b8b" (OuterVolumeSpecName: "kube-api-access-b6b8b") pod "432c1370-925d-4d16-9bc0-b7736387e058" (UID: "432c1370-925d-4d16-9bc0-b7736387e058"). InnerVolumeSpecName "kube-api-access-b6b8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.596240 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "74c3e024-0dfe-4f9c-930e-70896c3b8110" (UID: "74c3e024-0dfe-4f9c-930e-70896c3b8110"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.598383 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-kube-api-access-r6ck7" (OuterVolumeSpecName: "kube-api-access-r6ck7") pod "094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" (UID: "094d3d4c-b9d5-4d74-a851-7ac814d0f3e6"). InnerVolumeSpecName "kube-api-access-r6ck7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.598775 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c3e024-0dfe-4f9c-930e-70896c3b8110-kube-api-access-hjhww" (OuterVolumeSpecName: "kube-api-access-hjhww") pod "74c3e024-0dfe-4f9c-930e-70896c3b8110" (UID: "74c3e024-0dfe-4f9c-930e-70896c3b8110"). InnerVolumeSpecName "kube-api-access-hjhww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.622143 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "432c1370-925d-4d16-9bc0-b7736387e058" (UID: "432c1370-925d-4d16-9bc0-b7736387e058"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690306 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6ck7\" (UniqueName: \"kubernetes.io/projected/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-kube-api-access-r6ck7\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690528 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690591 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74c3e024-0dfe-4f9c-930e-70896c3b8110-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690650 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690717 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690786 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432c1370-925d-4d16-9bc0-b7736387e058-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690843 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6b8b\" (UniqueName: \"kubernetes.io/projected/432c1370-925d-4d16-9bc0-b7736387e058-kube-api-access-b6b8b\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.690904 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjhww\" (UniqueName: \"kubernetes.io/projected/74c3e024-0dfe-4f9c-930e-70896c3b8110-kube-api-access-hjhww\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.704198 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" (UID: "094d3d4c-b9d5-4d74-a851-7ac814d0f3e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.793538 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.831238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" event={"ID":"2f9ae171-32b7-4ca1-8d91-399067439899","Type":"ContainerStarted","Data":"d3ff51a5a3412bf456b03f817444bd51c1f29e341cbaac5e3a65308815acbc70"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.831306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" event={"ID":"2f9ae171-32b7-4ca1-8d91-399067439899","Type":"ContainerStarted","Data":"3693b03daf1b7cee7a69c72a8c48af92e0687dca48558d2f3a3f83688171f78f"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.832768 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.835531 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k4b69 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.835582 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" podUID="2f9ae171-32b7-4ca1-8d91-399067439899" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.836948 4810 generic.go:334] "Generic (PLEG): container finished" podID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerID="49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32" exitCode=0 Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.837089 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb4ww" event={"ID":"85ab7796-c612-4b27-a367-1d3deca20d7a","Type":"ContainerDied","Data":"49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.837128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb4ww" event={"ID":"85ab7796-c612-4b27-a367-1d3deca20d7a","Type":"ContainerDied","Data":"6e52cbe58263d14b93494349105f4c2d1cd5861513fb8137cfe14e13f5d810ac"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.837155 4810 scope.go:117] "RemoveContainer" containerID="49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.837241 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb4ww" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.846761 4810 generic.go:334] "Generic (PLEG): container finished" podID="432c1370-925d-4d16-9bc0-b7736387e058" containerID="9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd" exitCode=0 Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.846873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtjk2" event={"ID":"432c1370-925d-4d16-9bc0-b7736387e058","Type":"ContainerDied","Data":"9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.846910 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtjk2" event={"ID":"432c1370-925d-4d16-9bc0-b7736387e058","Type":"ContainerDied","Data":"e7537e535a039e14b923a7f1b47e78951de86ecf322f93e2eb410cea19ceb1ca"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.847059 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xtjk2" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.855352 4810 generic.go:334] "Generic (PLEG): container finished" podID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerID="7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" exitCode=0 Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.855444 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8xxs" event={"ID":"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6","Type":"ContainerDied","Data":"7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.855585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8xxs" event={"ID":"094d3d4c-b9d5-4d74-a851-7ac814d0f3e6","Type":"ContainerDied","Data":"d42498cfca0e06a790750719bfc9a41e9eee52d3e1f766682b4937f26092b28d"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.855678 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8xxs" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.862578 4810 generic.go:334] "Generic (PLEG): container finished" podID="74c3e024-0dfe-4f9c-930e-70896c3b8110" containerID="a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d" exitCode=0 Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.862656 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.862675 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" event={"ID":"74c3e024-0dfe-4f9c-930e-70896c3b8110","Type":"ContainerDied","Data":"a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.862704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7q29k" event={"ID":"74c3e024-0dfe-4f9c-930e-70896c3b8110","Type":"ContainerDied","Data":"7c1e338e4bee426d2b94f80afe5f4d7a6e66e0f6ca18319d1f30bfc23f746ffc"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.863241 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" podStartSLOduration=1.863228203 podStartE2EDuration="1.863228203s" podCreationTimestamp="2025-10-08 06:35:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:35:40.849126023 +0000 UTC m=+243.483565763" watchObservedRunningTime="2025-10-08 06:35:40.863228203 +0000 UTC m=+243.497667933" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.865727 4810 generic.go:334] "Generic (PLEG): container finished" podID="3abe3edb-3d3f-48a0-9968-f47012715666" containerID="aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe" exitCode=0 Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.865784 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fw7n" event={"ID":"3abe3edb-3d3f-48a0-9968-f47012715666","Type":"ContainerDied","Data":"aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.865811 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fw7n" event={"ID":"3abe3edb-3d3f-48a0-9968-f47012715666","Type":"ContainerDied","Data":"a8fb24019003f83f60f9f25fb3f2385cacad6f18c99ffe1d243c16542a9ae411"} Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.865902 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fw7n" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.871856 4810 scope.go:117] "RemoveContainer" containerID="989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.880632 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kb4ww"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.883551 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kb4ww"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.902552 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xtjk2"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.902634 4810 scope.go:117] "RemoveContainer" containerID="a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.905376 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xtjk2"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.922474 4810 scope.go:117] "RemoveContainer" containerID="49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.926326 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7fw7n"] Oct 08 06:35:40 crc kubenswrapper[4810]: E1008 06:35:40.928476 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32\": container with ID starting with 49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32 not found: ID does not exist" containerID="49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.928518 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32"} err="failed to get container status \"49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32\": rpc error: code = NotFound desc = could not find container \"49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32\": container with ID starting with 49b6707fd298eeac399c156eb162d211eed30d72430a3dee43fd3aeacd6f3d32 not found: ID does not exist" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.928553 4810 scope.go:117] "RemoveContainer" containerID="989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.928579 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7fw7n"] Oct 08 06:35:40 crc kubenswrapper[4810]: E1008 06:35:40.929233 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc\": container with ID starting with 989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc not found: ID does not exist" containerID="989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.929299 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc"} err="failed to get container status \"989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc\": rpc error: code = NotFound desc = could not find container \"989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc\": container with ID starting with 989369c7a7862682f759dc0913de143aefcb8087426df8113e5cfdd2885e22fc not found: ID does not exist" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.929348 4810 scope.go:117] "RemoveContainer" containerID="a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e" Oct 08 06:35:40 crc kubenswrapper[4810]: E1008 06:35:40.930093 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e\": container with ID starting with a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e not found: ID does not exist" containerID="a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.930133 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e"} err="failed to get container status \"a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e\": rpc error: code = NotFound desc = could not find container \"a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e\": container with ID starting with a4cf391f3d5e406ee76c37bc1cc8f0040bfd6786230b544ee7cdd77cfe41ed3e not found: ID does not exist" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.930155 4810 scope.go:117] "RemoveContainer" containerID="9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd" Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.936525 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7q29k"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.939692 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7q29k"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.945053 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8xxs"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.947987 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8xxs"] Oct 08 06:35:40 crc kubenswrapper[4810]: I1008 06:35:40.959745 4810 scope.go:117] "RemoveContainer" containerID="95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.009859 4810 scope.go:117] "RemoveContainer" containerID="100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.027817 4810 scope.go:117] "RemoveContainer" containerID="9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.028750 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd\": container with ID starting with 9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd not found: ID does not exist" containerID="9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.028799 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd"} err="failed to get container status \"9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd\": rpc error: code = NotFound desc = could not find container \"9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd\": container with ID starting with 9c7801a9224d2e4d82ff428d04c23d7152ede5da8d6d8cf40654cb26a15c6abd not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.028831 4810 scope.go:117] "RemoveContainer" containerID="95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.029220 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263\": container with ID starting with 95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263 not found: ID does not exist" containerID="95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.029263 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263"} err="failed to get container status \"95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263\": rpc error: code = NotFound desc = could not find container \"95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263\": container with ID starting with 95b037232e34a8b00a475ef7fea0b97d0427b401f7717772670b432aa005c263 not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.029296 4810 scope.go:117] "RemoveContainer" containerID="100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.029566 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151\": container with ID starting with 100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151 not found: ID does not exist" containerID="100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.029594 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151"} err="failed to get container status \"100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151\": rpc error: code = NotFound desc = could not find container \"100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151\": container with ID starting with 100eaf82cedf9dc1eefa3121ee22d1ff6da425a6d851fbaa9b72d85005be2151 not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.029611 4810 scope.go:117] "RemoveContainer" containerID="7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.042669 4810 scope.go:117] "RemoveContainer" containerID="da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.058652 4810 scope.go:117] "RemoveContainer" containerID="4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.075871 4810 scope.go:117] "RemoveContainer" containerID="7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.076250 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77\": container with ID starting with 7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77 not found: ID does not exist" containerID="7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.076284 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77"} err="failed to get container status \"7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77\": rpc error: code = NotFound desc = could not find container \"7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77\": container with ID starting with 7964cdc87c18ad676b155732802340bfa9cf29a85c1328be86801821829bce77 not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.076314 4810 scope.go:117] "RemoveContainer" containerID="da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.076580 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a\": container with ID starting with da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a not found: ID does not exist" containerID="da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.076604 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a"} err="failed to get container status \"da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a\": rpc error: code = NotFound desc = could not find container \"da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a\": container with ID starting with da3853f45c01c2eaf8081fa12381f83449b88d2fde5606f2ec91f62e4705738a not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.076618 4810 scope.go:117] "RemoveContainer" containerID="4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.076810 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c\": container with ID starting with 4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c not found: ID does not exist" containerID="4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.076829 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c"} err="failed to get container status \"4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c\": rpc error: code = NotFound desc = could not find container \"4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c\": container with ID starting with 4cdbf881c2830c5fc131c8a2e870380555f85cab175a3be6e28b0c8ca918567c not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.076841 4810 scope.go:117] "RemoveContainer" containerID="a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.089151 4810 scope.go:117] "RemoveContainer" containerID="a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.089500 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d\": container with ID starting with a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d not found: ID does not exist" containerID="a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.089526 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d"} err="failed to get container status \"a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d\": rpc error: code = NotFound desc = could not find container \"a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d\": container with ID starting with a4cb418eb5b09af8a1fb5f5f88f5594e6c9e6f7a45b20bc73f75fc02d4a1395d not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.089541 4810 scope.go:117] "RemoveContainer" containerID="aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.102341 4810 scope.go:117] "RemoveContainer" containerID="77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.120673 4810 scope.go:117] "RemoveContainer" containerID="cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.134260 4810 scope.go:117] "RemoveContainer" containerID="aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.135310 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe\": container with ID starting with aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe not found: ID does not exist" containerID="aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.135365 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe"} err="failed to get container status \"aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe\": rpc error: code = NotFound desc = could not find container \"aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe\": container with ID starting with aa73bbb5fd591f6005e59d628da600ae955a5feed7592dfeda121946294667fe not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.135406 4810 scope.go:117] "RemoveContainer" containerID="77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.135726 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6\": container with ID starting with 77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6 not found: ID does not exist" containerID="77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.135765 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6"} err="failed to get container status \"77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6\": rpc error: code = NotFound desc = could not find container \"77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6\": container with ID starting with 77d077c4934b1b7b58dc5650b8923308bad24cf07cfffc75e52d6c2c9ba951e6 not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.135789 4810 scope.go:117] "RemoveContainer" containerID="cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e" Oct 08 06:35:41 crc kubenswrapper[4810]: E1008 06:35:41.136099 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e\": container with ID starting with cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e not found: ID does not exist" containerID="cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.136147 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e"} err="failed to get container status \"cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e\": rpc error: code = NotFound desc = could not find container \"cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e\": container with ID starting with cd21381e28e57ae0e9b022630660833e3946d6e6e0a7ae5dbf290b7e8e944f7e not found: ID does not exist" Oct 08 06:35:41 crc kubenswrapper[4810]: I1008 06:35:41.888867 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-k4b69" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.034067 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zv7br"] Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035243 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035273 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035322 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035330 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035341 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035352 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035391 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c3e024-0dfe-4f9c-930e-70896c3b8110" containerName="marketplace-operator" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035399 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c3e024-0dfe-4f9c-930e-70896c3b8110" containerName="marketplace-operator" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035412 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035419 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035431 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035523 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035537 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035547 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035564 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035573 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035610 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035620 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035633 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035642 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035651 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035659 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035694 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035703 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="extract-content" Oct 08 06:35:42 crc kubenswrapper[4810]: E1008 06:35:42.035715 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035723 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="extract-utilities" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035890 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c3e024-0dfe-4f9c-930e-70896c3b8110" containerName="marketplace-operator" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035931 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035944 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="432c1370-925d-4d16-9bc0-b7736387e058" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.035998 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.036013 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" containerName="registry-server" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.037523 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.042642 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.047867 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zv7br"] Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.083569 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="094d3d4c-b9d5-4d74-a851-7ac814d0f3e6" path="/var/lib/kubelet/pods/094d3d4c-b9d5-4d74-a851-7ac814d0f3e6/volumes" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.084358 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3abe3edb-3d3f-48a0-9968-f47012715666" path="/var/lib/kubelet/pods/3abe3edb-3d3f-48a0-9968-f47012715666/volumes" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.085166 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="432c1370-925d-4d16-9bc0-b7736387e058" path="/var/lib/kubelet/pods/432c1370-925d-4d16-9bc0-b7736387e058/volumes" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.086493 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74c3e024-0dfe-4f9c-930e-70896c3b8110" path="/var/lib/kubelet/pods/74c3e024-0dfe-4f9c-930e-70896c3b8110/volumes" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.087101 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85ab7796-c612-4b27-a367-1d3deca20d7a" path="/var/lib/kubelet/pods/85ab7796-c612-4b27-a367-1d3deca20d7a/volumes" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.212835 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f658954e-7efb-4e24-95fa-b5191316ef0e-utilities\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.212884 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f658954e-7efb-4e24-95fa-b5191316ef0e-catalog-content\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.213081 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfpkw\" (UniqueName: \"kubernetes.io/projected/f658954e-7efb-4e24-95fa-b5191316ef0e-kube-api-access-rfpkw\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.235715 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jxmtt"] Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.236813 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.240198 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.252818 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jxmtt"] Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.314258 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfpkw\" (UniqueName: \"kubernetes.io/projected/f658954e-7efb-4e24-95fa-b5191316ef0e-kube-api-access-rfpkw\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.314364 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f658954e-7efb-4e24-95fa-b5191316ef0e-utilities\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.314390 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f658954e-7efb-4e24-95fa-b5191316ef0e-catalog-content\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.314921 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f658954e-7efb-4e24-95fa-b5191316ef0e-utilities\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.315032 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f658954e-7efb-4e24-95fa-b5191316ef0e-catalog-content\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.354376 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfpkw\" (UniqueName: \"kubernetes.io/projected/f658954e-7efb-4e24-95fa-b5191316ef0e-kube-api-access-rfpkw\") pod \"redhat-marketplace-zv7br\" (UID: \"f658954e-7efb-4e24-95fa-b5191316ef0e\") " pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.358321 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.415732 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0d2ff-103e-48be-a7e7-db75c60980f1-utilities\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.415790 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdwdp\" (UniqueName: \"kubernetes.io/projected/28a0d2ff-103e-48be-a7e7-db75c60980f1-kube-api-access-xdwdp\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.415827 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0d2ff-103e-48be-a7e7-db75c60980f1-catalog-content\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.517234 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdwdp\" (UniqueName: \"kubernetes.io/projected/28a0d2ff-103e-48be-a7e7-db75c60980f1-kube-api-access-xdwdp\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.517335 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0d2ff-103e-48be-a7e7-db75c60980f1-catalog-content\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.517409 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0d2ff-103e-48be-a7e7-db75c60980f1-utilities\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.518098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a0d2ff-103e-48be-a7e7-db75c60980f1-utilities\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.518818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a0d2ff-103e-48be-a7e7-db75c60980f1-catalog-content\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.545594 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdwdp\" (UniqueName: \"kubernetes.io/projected/28a0d2ff-103e-48be-a7e7-db75c60980f1-kube-api-access-xdwdp\") pod \"redhat-operators-jxmtt\" (UID: \"28a0d2ff-103e-48be-a7e7-db75c60980f1\") " pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.568064 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.625830 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zv7br"] Oct 08 06:35:42 crc kubenswrapper[4810]: W1008 06:35:42.637679 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf658954e_7efb_4e24_95fa_b5191316ef0e.slice/crio-381b8091a9e9fb458ac172a68599d25e92c1667a45d13a2be64a866ba8482fc0 WatchSource:0}: Error finding container 381b8091a9e9fb458ac172a68599d25e92c1667a45d13a2be64a866ba8482fc0: Status 404 returned error can't find the container with id 381b8091a9e9fb458ac172a68599d25e92c1667a45d13a2be64a866ba8482fc0 Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.897900 4810 generic.go:334] "Generic (PLEG): container finished" podID="f658954e-7efb-4e24-95fa-b5191316ef0e" containerID="abaaf4fb62b5d1a8992cfddd4d5cf5405f8fc318ed27a93a5ef9339579eb1a8e" exitCode=0 Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.898059 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zv7br" event={"ID":"f658954e-7efb-4e24-95fa-b5191316ef0e","Type":"ContainerDied","Data":"abaaf4fb62b5d1a8992cfddd4d5cf5405f8fc318ed27a93a5ef9339579eb1a8e"} Oct 08 06:35:42 crc kubenswrapper[4810]: I1008 06:35:42.898101 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zv7br" event={"ID":"f658954e-7efb-4e24-95fa-b5191316ef0e","Type":"ContainerStarted","Data":"381b8091a9e9fb458ac172a68599d25e92c1667a45d13a2be64a866ba8482fc0"} Oct 08 06:35:43 crc kubenswrapper[4810]: I1008 06:35:43.013893 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jxmtt"] Oct 08 06:35:43 crc kubenswrapper[4810]: W1008 06:35:43.018424 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28a0d2ff_103e_48be_a7e7_db75c60980f1.slice/crio-159b586073bc14487d1bab3703a2a9a42c02c937e233a8e8930cdc5b44e17062 WatchSource:0}: Error finding container 159b586073bc14487d1bab3703a2a9a42c02c937e233a8e8930cdc5b44e17062: Status 404 returned error can't find the container with id 159b586073bc14487d1bab3703a2a9a42c02c937e233a8e8930cdc5b44e17062 Oct 08 06:35:43 crc kubenswrapper[4810]: I1008 06:35:43.912025 4810 generic.go:334] "Generic (PLEG): container finished" podID="28a0d2ff-103e-48be-a7e7-db75c60980f1" containerID="e4ab28ad20c225e01a0aa9a03a8027c856044fdfbe15a96f4ffa6602261c010e" exitCode=0 Oct 08 06:35:43 crc kubenswrapper[4810]: I1008 06:35:43.912202 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxmtt" event={"ID":"28a0d2ff-103e-48be-a7e7-db75c60980f1","Type":"ContainerDied","Data":"e4ab28ad20c225e01a0aa9a03a8027c856044fdfbe15a96f4ffa6602261c010e"} Oct 08 06:35:43 crc kubenswrapper[4810]: I1008 06:35:43.912257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxmtt" event={"ID":"28a0d2ff-103e-48be-a7e7-db75c60980f1","Type":"ContainerStarted","Data":"159b586073bc14487d1bab3703a2a9a42c02c937e233a8e8930cdc5b44e17062"} Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.445802 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dl5kz"] Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.447950 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.452309 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.462457 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dl5kz"] Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.553609 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bblwg\" (UniqueName: \"kubernetes.io/projected/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-kube-api-access-bblwg\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.554387 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-catalog-content\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.554760 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-utilities\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.636403 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cv58h"] Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.638555 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.641558 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.645795 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cv58h"] Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.655954 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-utilities\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.656027 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bblwg\" (UniqueName: \"kubernetes.io/projected/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-kube-api-access-bblwg\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.656061 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-catalog-content\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.656580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-catalog-content\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.657287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-utilities\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.679259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bblwg\" (UniqueName: \"kubernetes.io/projected/2798b3c2-26d6-41e1-a3a4-ab10b5af84be-kube-api-access-bblwg\") pod \"community-operators-dl5kz\" (UID: \"2798b3c2-26d6-41e1-a3a4-ab10b5af84be\") " pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.757579 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb7qh\" (UniqueName: \"kubernetes.io/projected/90532ac2-5faa-4f71-b4ea-4c9af34d0199-kube-api-access-rb7qh\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.757694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90532ac2-5faa-4f71-b4ea-4c9af34d0199-utilities\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.757728 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90532ac2-5faa-4f71-b4ea-4c9af34d0199-catalog-content\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.781435 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.859166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90532ac2-5faa-4f71-b4ea-4c9af34d0199-utilities\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.859788 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90532ac2-5faa-4f71-b4ea-4c9af34d0199-catalog-content\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.859903 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb7qh\" (UniqueName: \"kubernetes.io/projected/90532ac2-5faa-4f71-b4ea-4c9af34d0199-kube-api-access-rb7qh\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.860059 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90532ac2-5faa-4f71-b4ea-4c9af34d0199-catalog-content\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.859689 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90532ac2-5faa-4f71-b4ea-4c9af34d0199-utilities\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.899781 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb7qh\" (UniqueName: \"kubernetes.io/projected/90532ac2-5faa-4f71-b4ea-4c9af34d0199-kube-api-access-rb7qh\") pod \"certified-operators-cv58h\" (UID: \"90532ac2-5faa-4f71-b4ea-4c9af34d0199\") " pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.924178 4810 generic.go:334] "Generic (PLEG): container finished" podID="f658954e-7efb-4e24-95fa-b5191316ef0e" containerID="d0bb33ac04decdb66480fa797d82885186579014ba3580f43618b9baf6173d47" exitCode=0 Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.924263 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zv7br" event={"ID":"f658954e-7efb-4e24-95fa-b5191316ef0e","Type":"ContainerDied","Data":"d0bb33ac04decdb66480fa797d82885186579014ba3580f43618b9baf6173d47"} Oct 08 06:35:44 crc kubenswrapper[4810]: I1008 06:35:44.930687 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxmtt" event={"ID":"28a0d2ff-103e-48be-a7e7-db75c60980f1","Type":"ContainerStarted","Data":"4b855c3da9e90a0b206bea409299dba48aec9026fce14fbab3b58b76c9621c66"} Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.029405 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.218852 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dl5kz"] Oct 08 06:35:45 crc kubenswrapper[4810]: W1008 06:35:45.226010 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2798b3c2_26d6_41e1_a3a4_ab10b5af84be.slice/crio-6fe5ffc6565d974ef28b213ad44d3e28c96cf4478117d566ece773956b98e7b8 WatchSource:0}: Error finding container 6fe5ffc6565d974ef28b213ad44d3e28c96cf4478117d566ece773956b98e7b8: Status 404 returned error can't find the container with id 6fe5ffc6565d974ef28b213ad44d3e28c96cf4478117d566ece773956b98e7b8 Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.440018 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cv58h"] Oct 08 06:35:45 crc kubenswrapper[4810]: W1008 06:35:45.465251 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90532ac2_5faa_4f71_b4ea_4c9af34d0199.slice/crio-505c6d6646664dc8e347403195bcae776526a597d6c7908474551558cec0aeee WatchSource:0}: Error finding container 505c6d6646664dc8e347403195bcae776526a597d6c7908474551558cec0aeee: Status 404 returned error can't find the container with id 505c6d6646664dc8e347403195bcae776526a597d6c7908474551558cec0aeee Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.952098 4810 generic.go:334] "Generic (PLEG): container finished" podID="90532ac2-5faa-4f71-b4ea-4c9af34d0199" containerID="7e0aa4ecd8896dff88558c047ff909066e32552e22c33cfa65140fac465411c4" exitCode=0 Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.952208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cv58h" event={"ID":"90532ac2-5faa-4f71-b4ea-4c9af34d0199","Type":"ContainerDied","Data":"7e0aa4ecd8896dff88558c047ff909066e32552e22c33cfa65140fac465411c4"} Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.954194 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cv58h" event={"ID":"90532ac2-5faa-4f71-b4ea-4c9af34d0199","Type":"ContainerStarted","Data":"505c6d6646664dc8e347403195bcae776526a597d6c7908474551558cec0aeee"} Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.962081 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zv7br" event={"ID":"f658954e-7efb-4e24-95fa-b5191316ef0e","Type":"ContainerStarted","Data":"d146cc163b91afe450dbf6ad318aa4b6ed15f4818e9733a31a66b9ea13a32fa7"} Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.963802 4810 generic.go:334] "Generic (PLEG): container finished" podID="28a0d2ff-103e-48be-a7e7-db75c60980f1" containerID="4b855c3da9e90a0b206bea409299dba48aec9026fce14fbab3b58b76c9621c66" exitCode=0 Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.963884 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxmtt" event={"ID":"28a0d2ff-103e-48be-a7e7-db75c60980f1","Type":"ContainerDied","Data":"4b855c3da9e90a0b206bea409299dba48aec9026fce14fbab3b58b76c9621c66"} Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.972570 4810 generic.go:334] "Generic (PLEG): container finished" podID="2798b3c2-26d6-41e1-a3a4-ab10b5af84be" containerID="dd41beebf089af683e7886dd50f81256cc9dac6163853173f1f081cd62b723f1" exitCode=0 Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.972626 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl5kz" event={"ID":"2798b3c2-26d6-41e1-a3a4-ab10b5af84be","Type":"ContainerDied","Data":"dd41beebf089af683e7886dd50f81256cc9dac6163853173f1f081cd62b723f1"} Oct 08 06:35:45 crc kubenswrapper[4810]: I1008 06:35:45.972661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl5kz" event={"ID":"2798b3c2-26d6-41e1-a3a4-ab10b5af84be","Type":"ContainerStarted","Data":"6fe5ffc6565d974ef28b213ad44d3e28c96cf4478117d566ece773956b98e7b8"} Oct 08 06:35:46 crc kubenswrapper[4810]: I1008 06:35:46.020174 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zv7br" podStartSLOduration=1.537776372 podStartE2EDuration="4.020153841s" podCreationTimestamp="2025-10-08 06:35:42 +0000 UTC" firstStartedPulling="2025-10-08 06:35:42.90100815 +0000 UTC m=+245.535447890" lastFinishedPulling="2025-10-08 06:35:45.383385619 +0000 UTC m=+248.017825359" observedRunningTime="2025-10-08 06:35:46.017070451 +0000 UTC m=+248.651510191" watchObservedRunningTime="2025-10-08 06:35:46.020153841 +0000 UTC m=+248.654593581" Oct 08 06:35:46 crc kubenswrapper[4810]: I1008 06:35:46.983071 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxmtt" event={"ID":"28a0d2ff-103e-48be-a7e7-db75c60980f1","Type":"ContainerStarted","Data":"a7073d46cfd505b6e60891f0b91562972d425e8ce3ef1f0223b3a4608aeed680"} Oct 08 06:35:46 crc kubenswrapper[4810]: I1008 06:35:46.986051 4810 generic.go:334] "Generic (PLEG): container finished" podID="2798b3c2-26d6-41e1-a3a4-ab10b5af84be" containerID="9731da92b3915e5e68861943ecbc532a888607f1109574b91006bc696aec74b4" exitCode=0 Oct 08 06:35:46 crc kubenswrapper[4810]: I1008 06:35:46.986112 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl5kz" event={"ID":"2798b3c2-26d6-41e1-a3a4-ab10b5af84be","Type":"ContainerDied","Data":"9731da92b3915e5e68861943ecbc532a888607f1109574b91006bc696aec74b4"} Oct 08 06:35:46 crc kubenswrapper[4810]: I1008 06:35:46.988644 4810 generic.go:334] "Generic (PLEG): container finished" podID="90532ac2-5faa-4f71-b4ea-4c9af34d0199" containerID="8d8a153a33baa3a0d4986d80627ac975e4f552f8710bdb771548d4c7f04ac1dc" exitCode=0 Oct 08 06:35:46 crc kubenswrapper[4810]: I1008 06:35:46.988762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cv58h" event={"ID":"90532ac2-5faa-4f71-b4ea-4c9af34d0199","Type":"ContainerDied","Data":"8d8a153a33baa3a0d4986d80627ac975e4f552f8710bdb771548d4c7f04ac1dc"} Oct 08 06:35:47 crc kubenswrapper[4810]: I1008 06:35:47.007897 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jxmtt" podStartSLOduration=2.525148534 podStartE2EDuration="5.007876263s" podCreationTimestamp="2025-10-08 06:35:42 +0000 UTC" firstStartedPulling="2025-10-08 06:35:43.918377283 +0000 UTC m=+246.552817023" lastFinishedPulling="2025-10-08 06:35:46.401105012 +0000 UTC m=+249.035544752" observedRunningTime="2025-10-08 06:35:47.00400933 +0000 UTC m=+249.638449090" watchObservedRunningTime="2025-10-08 06:35:47.007876263 +0000 UTC m=+249.642316013" Oct 08 06:35:49 crc kubenswrapper[4810]: I1008 06:35:49.016682 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dl5kz" event={"ID":"2798b3c2-26d6-41e1-a3a4-ab10b5af84be","Type":"ContainerStarted","Data":"43ed7ac6b9d80d7ce4abc20806af811cc625993aa60799c99bb86503f258d6e4"} Oct 08 06:35:49 crc kubenswrapper[4810]: I1008 06:35:49.027648 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cv58h" event={"ID":"90532ac2-5faa-4f71-b4ea-4c9af34d0199","Type":"ContainerStarted","Data":"869316d6b9def87b0ce9732305c8633f0077a7aa9f95de17e202dada3843c619"} Oct 08 06:35:49 crc kubenswrapper[4810]: I1008 06:35:49.053701 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dl5kz" podStartSLOduration=3.580193161 podStartE2EDuration="5.053675942s" podCreationTimestamp="2025-10-08 06:35:44 +0000 UTC" firstStartedPulling="2025-10-08 06:35:45.975527833 +0000 UTC m=+248.609967563" lastFinishedPulling="2025-10-08 06:35:47.449010614 +0000 UTC m=+250.083450344" observedRunningTime="2025-10-08 06:35:49.051246601 +0000 UTC m=+251.685686341" watchObservedRunningTime="2025-10-08 06:35:49.053675942 +0000 UTC m=+251.688115682" Oct 08 06:35:49 crc kubenswrapper[4810]: I1008 06:35:49.086131 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cv58h" podStartSLOduration=3.354848885 podStartE2EDuration="5.086110795s" podCreationTimestamp="2025-10-08 06:35:44 +0000 UTC" firstStartedPulling="2025-10-08 06:35:45.956855759 +0000 UTC m=+248.591295499" lastFinishedPulling="2025-10-08 06:35:47.688117669 +0000 UTC m=+250.322557409" observedRunningTime="2025-10-08 06:35:49.082682255 +0000 UTC m=+251.717121995" watchObservedRunningTime="2025-10-08 06:35:49.086110795 +0000 UTC m=+251.720550535" Oct 08 06:35:52 crc kubenswrapper[4810]: I1008 06:35:52.359437 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:52 crc kubenswrapper[4810]: I1008 06:35:52.362180 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:52 crc kubenswrapper[4810]: I1008 06:35:52.404243 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:52 crc kubenswrapper[4810]: I1008 06:35:52.569609 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:52 crc kubenswrapper[4810]: I1008 06:35:52.569659 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:52 crc kubenswrapper[4810]: I1008 06:35:52.620080 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:53 crc kubenswrapper[4810]: I1008 06:35:53.105389 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zv7br" Oct 08 06:35:53 crc kubenswrapper[4810]: I1008 06:35:53.106172 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jxmtt" Oct 08 06:35:54 crc kubenswrapper[4810]: I1008 06:35:54.782745 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:54 crc kubenswrapper[4810]: I1008 06:35:54.783066 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:54 crc kubenswrapper[4810]: I1008 06:35:54.849310 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:55 crc kubenswrapper[4810]: I1008 06:35:55.031714 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:55 crc kubenswrapper[4810]: I1008 06:35:55.031799 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:55 crc kubenswrapper[4810]: I1008 06:35:55.074120 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:35:55 crc kubenswrapper[4810]: I1008 06:35:55.122226 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dl5kz" Oct 08 06:35:55 crc kubenswrapper[4810]: I1008 06:35:55.144948 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cv58h" Oct 08 06:37:19 crc kubenswrapper[4810]: I1008 06:37:19.432765 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:37:19 crc kubenswrapper[4810]: I1008 06:37:19.434233 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:37:49 crc kubenswrapper[4810]: I1008 06:37:49.432437 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:37:49 crc kubenswrapper[4810]: I1008 06:37:49.434183 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:38:19 crc kubenswrapper[4810]: I1008 06:38:19.433109 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:38:19 crc kubenswrapper[4810]: I1008 06:38:19.433953 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:38:19 crc kubenswrapper[4810]: I1008 06:38:19.434082 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:38:19 crc kubenswrapper[4810]: I1008 06:38:19.434957 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fba6723b31f4f6438b0e473e38431e96175c272a5ec8aa181205f38fc5e27943"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:38:19 crc kubenswrapper[4810]: I1008 06:38:19.435103 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://fba6723b31f4f6438b0e473e38431e96175c272a5ec8aa181205f38fc5e27943" gracePeriod=600 Oct 08 06:38:20 crc kubenswrapper[4810]: I1008 06:38:20.178451 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="fba6723b31f4f6438b0e473e38431e96175c272a5ec8aa181205f38fc5e27943" exitCode=0 Oct 08 06:38:20 crc kubenswrapper[4810]: I1008 06:38:20.178508 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"fba6723b31f4f6438b0e473e38431e96175c272a5ec8aa181205f38fc5e27943"} Oct 08 06:38:20 crc kubenswrapper[4810]: I1008 06:38:20.178949 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"f85489310335fb08e0e9aeb0f5009d545bd1d41a5c6acf3db9afeeed7e8e5528"} Oct 08 06:38:20 crc kubenswrapper[4810]: I1008 06:38:20.179022 4810 scope.go:117] "RemoveContainer" containerID="ced56b1fab796906c3ea50e0d1bd27cf399fe96ab93b3fbb9eb1bccf21b9feea" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.276299 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kz2n5"] Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.278177 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.289841 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kz2n5"] Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.388835 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-registry-certificates\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.388917 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.388957 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-bound-sa-token\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.389007 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-trusted-ca\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.389035 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-registry-tls\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.389059 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.389291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.389338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lktbq\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-kube-api-access-lktbq\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.415598 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.491183 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-registry-certificates\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.491236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-bound-sa-token\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.491263 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-trusted-ca\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.491283 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-registry-tls\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.491300 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.491333 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.491362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lktbq\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-kube-api-access-lktbq\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.492349 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.493435 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-registry-certificates\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.494445 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-trusted-ca\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.501701 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.502237 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-registry-tls\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.517624 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-bound-sa-token\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.519158 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lktbq\" (UniqueName: \"kubernetes.io/projected/2ff22bc3-f4b7-47a8-94a7-8af77c4661e3-kube-api-access-lktbq\") pod \"image-registry-66df7c8f76-kz2n5\" (UID: \"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3\") " pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.602634 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:35 crc kubenswrapper[4810]: I1008 06:38:35.890603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-kz2n5"] Oct 08 06:38:35 crc kubenswrapper[4810]: W1008 06:38:35.893561 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ff22bc3_f4b7_47a8_94a7_8af77c4661e3.slice/crio-e81c3f84d204163b56c37d724164cf3a699bfaaacd97e9302aec71a5e7f54fec WatchSource:0}: Error finding container e81c3f84d204163b56c37d724164cf3a699bfaaacd97e9302aec71a5e7f54fec: Status 404 returned error can't find the container with id e81c3f84d204163b56c37d724164cf3a699bfaaacd97e9302aec71a5e7f54fec Oct 08 06:38:36 crc kubenswrapper[4810]: I1008 06:38:36.315531 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" event={"ID":"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3","Type":"ContainerStarted","Data":"5a9d9b8ff256d6d69b10bba2f077ba81d9e1466c453757a32af7b6f63625eaf3"} Oct 08 06:38:36 crc kubenswrapper[4810]: I1008 06:38:36.315632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" event={"ID":"2ff22bc3-f4b7-47a8-94a7-8af77c4661e3","Type":"ContainerStarted","Data":"e81c3f84d204163b56c37d724164cf3a699bfaaacd97e9302aec71a5e7f54fec"} Oct 08 06:38:36 crc kubenswrapper[4810]: I1008 06:38:36.315814 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:36 crc kubenswrapper[4810]: I1008 06:38:36.341632 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" podStartSLOduration=1.341574138 podStartE2EDuration="1.341574138s" podCreationTimestamp="2025-10-08 06:38:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:38:36.336739856 +0000 UTC m=+418.971179636" watchObservedRunningTime="2025-10-08 06:38:36.341574138 +0000 UTC m=+418.976013898" Oct 08 06:38:55 crc kubenswrapper[4810]: I1008 06:38:55.614307 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-kz2n5" Oct 08 06:38:55 crc kubenswrapper[4810]: I1008 06:38:55.689883 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gpgrc"] Oct 08 06:39:20 crc kubenswrapper[4810]: I1008 06:39:20.783049 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" podUID="a714d476-65eb-4e18-a0ba-e874bae2ca79" containerName="registry" containerID="cri-o://2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044" gracePeriod=30 Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.168814 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210173 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-tls\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210326 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a714d476-65eb-4e18-a0ba-e874bae2ca79-installation-pull-secrets\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210359 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-certificates\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210386 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-trusted-ca\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210611 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210694 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbz8m\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-kube-api-access-vbz8m\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210753 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a714d476-65eb-4e18-a0ba-e874bae2ca79-ca-trust-extracted\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.210778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-bound-sa-token\") pod \"a714d476-65eb-4e18-a0ba-e874bae2ca79\" (UID: \"a714d476-65eb-4e18-a0ba-e874bae2ca79\") " Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.212527 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.212792 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.217989 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-kube-api-access-vbz8m" (OuterVolumeSpecName: "kube-api-access-vbz8m") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "kube-api-access-vbz8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.219281 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.220469 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a714d476-65eb-4e18-a0ba-e874bae2ca79-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.220618 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.227638 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.240689 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a714d476-65eb-4e18-a0ba-e874bae2ca79-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a714d476-65eb-4e18-a0ba-e874bae2ca79" (UID: "a714d476-65eb-4e18-a0ba-e874bae2ca79"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.313280 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbz8m\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-kube-api-access-vbz8m\") on node \"crc\" DevicePath \"\"" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.314141 4810 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a714d476-65eb-4e18-a0ba-e874bae2ca79-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.314158 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.314175 4810 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.314215 4810 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a714d476-65eb-4e18-a0ba-e874bae2ca79-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.314227 4810 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.314238 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a714d476-65eb-4e18-a0ba-e874bae2ca79-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.632015 4810 generic.go:334] "Generic (PLEG): container finished" podID="a714d476-65eb-4e18-a0ba-e874bae2ca79" containerID="2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044" exitCode=0 Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.632088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" event={"ID":"a714d476-65eb-4e18-a0ba-e874bae2ca79","Type":"ContainerDied","Data":"2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044"} Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.632121 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" event={"ID":"a714d476-65eb-4e18-a0ba-e874bae2ca79","Type":"ContainerDied","Data":"4e1597d5658f0563aedccc5e43ed2447e78dbf8639a010a134597380cc78a79d"} Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.632139 4810 scope.go:117] "RemoveContainer" containerID="2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.632151 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gpgrc" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.664465 4810 scope.go:117] "RemoveContainer" containerID="2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044" Oct 08 06:39:21 crc kubenswrapper[4810]: E1008 06:39:21.665469 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044\": container with ID starting with 2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044 not found: ID does not exist" containerID="2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.665589 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044"} err="failed to get container status \"2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044\": rpc error: code = NotFound desc = could not find container \"2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044\": container with ID starting with 2f27226255b7bcb8650581db5dde83b1342c58e780a0878d8e95fc45dee8c044 not found: ID does not exist" Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.696899 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gpgrc"] Oct 08 06:39:21 crc kubenswrapper[4810]: I1008 06:39:21.702676 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gpgrc"] Oct 08 06:39:22 crc kubenswrapper[4810]: I1008 06:39:22.092835 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a714d476-65eb-4e18-a0ba-e874bae2ca79" path="/var/lib/kubelet/pods/a714d476-65eb-4e18-a0ba-e874bae2ca79/volumes" Oct 08 06:40:19 crc kubenswrapper[4810]: I1008 06:40:19.431641 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:40:19 crc kubenswrapper[4810]: I1008 06:40:19.432487 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:40:49 crc kubenswrapper[4810]: I1008 06:40:49.432472 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:40:49 crc kubenswrapper[4810]: I1008 06:40:49.433289 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:41:19 crc kubenswrapper[4810]: I1008 06:41:19.431254 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:41:19 crc kubenswrapper[4810]: I1008 06:41:19.431824 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:41:19 crc kubenswrapper[4810]: I1008 06:41:19.431899 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:41:19 crc kubenswrapper[4810]: I1008 06:41:19.432839 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f85489310335fb08e0e9aeb0f5009d545bd1d41a5c6acf3db9afeeed7e8e5528"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:41:19 crc kubenswrapper[4810]: I1008 06:41:19.432941 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://f85489310335fb08e0e9aeb0f5009d545bd1d41a5c6acf3db9afeeed7e8e5528" gracePeriod=600 Oct 08 06:41:20 crc kubenswrapper[4810]: I1008 06:41:20.453129 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="f85489310335fb08e0e9aeb0f5009d545bd1d41a5c6acf3db9afeeed7e8e5528" exitCode=0 Oct 08 06:41:20 crc kubenswrapper[4810]: I1008 06:41:20.453178 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"f85489310335fb08e0e9aeb0f5009d545bd1d41a5c6acf3db9afeeed7e8e5528"} Oct 08 06:41:20 crc kubenswrapper[4810]: I1008 06:41:20.453717 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"2a91a7b94b569323b6b5755a125c6be2ca2114d957f75fbc9389bdbf60792809"} Oct 08 06:41:20 crc kubenswrapper[4810]: I1008 06:41:20.453749 4810 scope.go:117] "RemoveContainer" containerID="fba6723b31f4f6438b0e473e38431e96175c272a5ec8aa181205f38fc5e27943" Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.906820 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkrh9"] Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.908083 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-controller" containerID="cri-o://e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" gracePeriod=30 Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.908259 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-node" containerID="cri-o://12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" gracePeriod=30 Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.908201 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="nbdb" containerID="cri-o://c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" gracePeriod=30 Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.908348 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="northd" containerID="cri-o://6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" gracePeriod=30 Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.908332 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-acl-logging" containerID="cri-o://2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" gracePeriod=30 Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.908561 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="sbdb" containerID="cri-o://8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" gracePeriod=30 Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.908236 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" gracePeriod=30 Oct 08 06:41:48 crc kubenswrapper[4810]: I1008 06:41:48.951018 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" containerID="cri-o://f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" gracePeriod=30 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.249517 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/3.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.252451 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovn-acl-logging/0.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.252935 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovn-controller/0.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.253359 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.303547 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4b59c"] Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.303856 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a714d476-65eb-4e18-a0ba-e874bae2ca79" containerName="registry" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.303875 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a714d476-65eb-4e18-a0ba-e874bae2ca79" containerName="registry" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.303886 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kubecfg-setup" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.303894 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kubecfg-setup" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.303910 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="nbdb" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.303920 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="nbdb" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.303928 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="sbdb" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.303935 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="sbdb" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.303944 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.303951 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.303998 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304010 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304024 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-acl-logging" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304035 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-acl-logging" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304045 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304053 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304064 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304083 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304093 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304100 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304113 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="northd" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304122 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="northd" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304132 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-node" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304140 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-node" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304271 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a714d476-65eb-4e18-a0ba-e874bae2ca79" containerName="registry" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304287 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304299 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304310 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304321 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304336 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="sbdb" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304348 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304361 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="kube-rbac-proxy-node" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304371 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="nbdb" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304382 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-acl-logging" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304392 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="northd" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304402 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovn-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304515 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304525 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304703 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.304825 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.304839 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerName="ovnkube-controller" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.306325 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329103 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-systemd\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329145 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-ovn-kubernetes\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329172 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-var-lib-openvswitch\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329192 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-bin\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329213 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-openvswitch\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329230 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-systemd-units\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329245 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-netd\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329265 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329308 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329330 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329285 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-env-overrides\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329275 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329395 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-kubelet\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329360 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329446 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hzkl\" (UniqueName: \"kubernetes.io/projected/ee7b17fc-599e-4723-80fb-eb248dfe539d-kube-api-access-6hzkl\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329428 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329550 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-ovn\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329569 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-netns\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329492 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329552 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329599 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-log-socket\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-script-lib\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329668 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-slash\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329707 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-config\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329726 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-etc-openvswitch\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329605 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329646 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-log-socket" (OuterVolumeSpecName: "log-socket") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329746 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-slash" (OuterVolumeSpecName: "host-slash") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329751 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovn-node-metrics-cert\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.329809 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-node-log\") pod \"ee7b17fc-599e-4723-80fb-eb248dfe539d\" (UID: \"ee7b17fc-599e-4723-80fb-eb248dfe539d\") " Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330011 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-systemd\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330041 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330074 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-cni-bin\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330109 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-systemd-units\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330131 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-ovnkube-config\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330141 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330150 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330184 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-cni-netd\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330216 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-node-log" (OuterVolumeSpecName: "node-log") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330234 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-ovn\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330393 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-var-lib-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330478 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccgj8\" (UniqueName: \"kubernetes.io/projected/5f8702df-efc2-4949-b036-fc8433cc98ea-kube-api-access-ccgj8\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330496 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-log-socket\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330521 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-run-netns\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330544 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-ovnkube-script-lib\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330567 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-kubelet\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330587 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f8702df-efc2-4949-b036-fc8433cc98ea-ovn-node-metrics-cert\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330644 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330681 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-slash\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330700 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-node-log\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330717 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-etc-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330739 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-env-overrides\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330906 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330922 4810 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330935 4810 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330947 4810 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330957 4810 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330985 4810 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-log-socket\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.330999 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331009 4810 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-slash\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331019 4810 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331028 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331037 4810 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-node-log\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331046 4810 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331055 4810 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331063 4810 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331072 4810 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331081 4810 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.331089 4810 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.340690 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.340229 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee7b17fc-599e-4723-80fb-eb248dfe539d-kube-api-access-6hzkl" (OuterVolumeSpecName: "kube-api-access-6hzkl") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "kube-api-access-6hzkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.350799 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ee7b17fc-599e-4723-80fb-eb248dfe539d" (UID: "ee7b17fc-599e-4723-80fb-eb248dfe539d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.431945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccgj8\" (UniqueName: \"kubernetes.io/projected/5f8702df-efc2-4949-b036-fc8433cc98ea-kube-api-access-ccgj8\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432087 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-log-socket\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432121 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-run-netns\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432151 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-ovnkube-script-lib\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432175 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-kubelet\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f8702df-efc2-4949-b036-fc8433cc98ea-ovn-node-metrics-cert\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432240 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432265 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-slash\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432310 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-node-log\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432324 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-log-socket\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432336 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-kubelet\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-etc-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-etc-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432443 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-env-overrides\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432468 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432491 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-systemd\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432507 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-slash\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432544 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-node-log\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432559 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-cni-bin\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432589 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-systemd-units\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432620 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-ovnkube-config\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432661 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-cni-netd\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432689 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-ovn\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432752 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-cni-bin\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432799 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-var-lib-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-var-lib-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.432850 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-cni-netd\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433024 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-systemd\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433075 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-ovn\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433103 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-systemd-units\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-run-openvswitch\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f8702df-efc2-4949-b036-fc8433cc98ea-host-run-netns\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433205 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hzkl\" (UniqueName: \"kubernetes.io/projected/ee7b17fc-599e-4723-80fb-eb248dfe539d-kube-api-access-6hzkl\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433228 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ee7b17fc-599e-4723-80fb-eb248dfe539d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433250 4810 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ee7b17fc-599e-4723-80fb-eb248dfe539d-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433531 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-ovnkube-config\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433535 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-ovnkube-script-lib\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.433669 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f8702df-efc2-4949-b036-fc8433cc98ea-env-overrides\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.437042 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f8702df-efc2-4949-b036-fc8433cc98ea-ovn-node-metrics-cert\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.448668 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccgj8\" (UniqueName: \"kubernetes.io/projected/5f8702df-efc2-4949-b036-fc8433cc98ea-kube-api-access-ccgj8\") pod \"ovnkube-node-4b59c\" (UID: \"5f8702df-efc2-4949-b036-fc8433cc98ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.623666 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.627727 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/2.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.628418 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/1.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.628505 4810 generic.go:334] "Generic (PLEG): container finished" podID="9f654c73-c1a2-4292-abe0-830b46ed68f5" containerID="97724d3ceaf0c93d52284cdd2d590748da432288355b5517444c117dcb5b06d9" exitCode=2 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.628635 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerDied","Data":"97724d3ceaf0c93d52284cdd2d590748da432288355b5517444c117dcb5b06d9"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.628735 4810 scope.go:117] "RemoveContainer" containerID="2841267f2e5ea0e78c31f32217f97ecb5c615b2c195b3bc140bb7b6cf778b6c1" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.629194 4810 scope.go:117] "RemoveContainer" containerID="97724d3ceaf0c93d52284cdd2d590748da432288355b5517444c117dcb5b06d9" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.629374 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-txvr6_openshift-multus(9f654c73-c1a2-4292-abe0-830b46ed68f5)\"" pod="openshift-multus/multus-txvr6" podUID="9f654c73-c1a2-4292-abe0-830b46ed68f5" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.631282 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovnkube-controller/3.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.635247 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovn-acl-logging/0.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.636747 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkrh9_ee7b17fc-599e-4723-80fb-eb248dfe539d/ovn-controller/0.log" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637629 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" exitCode=0 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637688 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" exitCode=0 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637704 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" exitCode=0 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637720 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" exitCode=0 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637735 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" exitCode=0 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637803 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" exitCode=0 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637821 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" exitCode=143 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637837 4810 generic.go:334] "Generic (PLEG): container finished" podID="ee7b17fc-599e-4723-80fb-eb248dfe539d" containerID="e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" exitCode=143 Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637803 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.637932 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638021 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638059 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638112 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638139 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638165 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638192 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638208 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638222 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638236 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638251 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638267 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638282 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638296 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638310 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638354 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638371 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638385 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638398 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638412 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638426 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638440 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638454 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638468 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638481 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638500 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638525 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638540 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638554 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638771 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.638995 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639018 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639031 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639047 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639061 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639074 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkrh9" event={"ID":"ee7b17fc-599e-4723-80fb-eb248dfe539d","Type":"ContainerDied","Data":"ee0a6f39b0a4e372e72a213ffb6f6ad0a5266d45d7ea3f803f14dcf33a8b2539"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639264 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639285 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639300 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639314 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639444 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639467 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639481 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639497 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639511 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.639597 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.669312 4810 scope.go:117] "RemoveContainer" containerID="f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.707298 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.711941 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkrh9"] Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.718315 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkrh9"] Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.759216 4810 scope.go:117] "RemoveContainer" containerID="8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.793289 4810 scope.go:117] "RemoveContainer" containerID="c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.808153 4810 scope.go:117] "RemoveContainer" containerID="6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.823832 4810 scope.go:117] "RemoveContainer" containerID="b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.916453 4810 scope.go:117] "RemoveContainer" containerID="12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.941901 4810 scope.go:117] "RemoveContainer" containerID="2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.963244 4810 scope.go:117] "RemoveContainer" containerID="e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.979209 4810 scope.go:117] "RemoveContainer" containerID="8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.991351 4810 scope.go:117] "RemoveContainer" containerID="f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.991720 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": container with ID starting with f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412 not found: ID does not exist" containerID="f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.991796 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} err="failed to get container status \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": rpc error: code = NotFound desc = could not find container \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": container with ID starting with f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.991828 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.992253 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": container with ID starting with b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5 not found: ID does not exist" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.992312 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} err="failed to get container status \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": rpc error: code = NotFound desc = could not find container \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": container with ID starting with b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.992344 4810 scope.go:117] "RemoveContainer" containerID="8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.992645 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": container with ID starting with 8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21 not found: ID does not exist" containerID="8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.992667 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} err="failed to get container status \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": rpc error: code = NotFound desc = could not find container \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": container with ID starting with 8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.992679 4810 scope.go:117] "RemoveContainer" containerID="c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.992939 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": container with ID starting with c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546 not found: ID does not exist" containerID="c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.992958 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} err="failed to get container status \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": rpc error: code = NotFound desc = could not find container \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": container with ID starting with c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.992991 4810 scope.go:117] "RemoveContainer" containerID="6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.993531 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": container with ID starting with 6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3 not found: ID does not exist" containerID="6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.993566 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} err="failed to get container status \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": rpc error: code = NotFound desc = could not find container \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": container with ID starting with 6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.993590 4810 scope.go:117] "RemoveContainer" containerID="b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.993842 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": container with ID starting with b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4 not found: ID does not exist" containerID="b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.993888 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} err="failed to get container status \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": rpc error: code = NotFound desc = could not find container \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": container with ID starting with b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.993958 4810 scope.go:117] "RemoveContainer" containerID="12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.997188 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": container with ID starting with 12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45 not found: ID does not exist" containerID="12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.997208 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} err="failed to get container status \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": rpc error: code = NotFound desc = could not find container \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": container with ID starting with 12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.997225 4810 scope.go:117] "RemoveContainer" containerID="2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.997480 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": container with ID starting with 2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862 not found: ID does not exist" containerID="2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.997511 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} err="failed to get container status \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": rpc error: code = NotFound desc = could not find container \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": container with ID starting with 2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.997531 4810 scope.go:117] "RemoveContainer" containerID="e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.997733 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": container with ID starting with e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd not found: ID does not exist" containerID="e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.997764 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} err="failed to get container status \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": rpc error: code = NotFound desc = could not find container \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": container with ID starting with e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.997778 4810 scope.go:117] "RemoveContainer" containerID="8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41" Oct 08 06:41:49 crc kubenswrapper[4810]: E1008 06:41:49.998023 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": container with ID starting with 8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41 not found: ID does not exist" containerID="8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998048 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} err="failed to get container status \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": rpc error: code = NotFound desc = could not find container \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": container with ID starting with 8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998065 4810 scope.go:117] "RemoveContainer" containerID="f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998245 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} err="failed to get container status \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": rpc error: code = NotFound desc = could not find container \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": container with ID starting with f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998264 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998443 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} err="failed to get container status \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": rpc error: code = NotFound desc = could not find container \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": container with ID starting with b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998460 4810 scope.go:117] "RemoveContainer" containerID="8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998641 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} err="failed to get container status \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": rpc error: code = NotFound desc = could not find container \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": container with ID starting with 8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.998660 4810 scope.go:117] "RemoveContainer" containerID="c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.999660 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} err="failed to get container status \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": rpc error: code = NotFound desc = could not find container \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": container with ID starting with c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.999690 4810 scope.go:117] "RemoveContainer" containerID="6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.999898 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} err="failed to get container status \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": rpc error: code = NotFound desc = could not find container \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": container with ID starting with 6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3 not found: ID does not exist" Oct 08 06:41:49 crc kubenswrapper[4810]: I1008 06:41:49.999931 4810 scope.go:117] "RemoveContainer" containerID="b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.000196 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} err="failed to get container status \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": rpc error: code = NotFound desc = could not find container \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": container with ID starting with b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.000248 4810 scope.go:117] "RemoveContainer" containerID="12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.000482 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} err="failed to get container status \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": rpc error: code = NotFound desc = could not find container \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": container with ID starting with 12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.000505 4810 scope.go:117] "RemoveContainer" containerID="2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.000770 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} err="failed to get container status \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": rpc error: code = NotFound desc = could not find container \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": container with ID starting with 2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.000793 4810 scope.go:117] "RemoveContainer" containerID="e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.001180 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} err="failed to get container status \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": rpc error: code = NotFound desc = could not find container \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": container with ID starting with e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.001218 4810 scope.go:117] "RemoveContainer" containerID="8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.001907 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} err="failed to get container status \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": rpc error: code = NotFound desc = could not find container \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": container with ID starting with 8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.001958 4810 scope.go:117] "RemoveContainer" containerID="f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.002492 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} err="failed to get container status \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": rpc error: code = NotFound desc = could not find container \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": container with ID starting with f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.002525 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.003238 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} err="failed to get container status \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": rpc error: code = NotFound desc = could not find container \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": container with ID starting with b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.003261 4810 scope.go:117] "RemoveContainer" containerID="8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.003602 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} err="failed to get container status \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": rpc error: code = NotFound desc = could not find container \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": container with ID starting with 8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.003620 4810 scope.go:117] "RemoveContainer" containerID="c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.003997 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} err="failed to get container status \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": rpc error: code = NotFound desc = could not find container \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": container with ID starting with c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.004091 4810 scope.go:117] "RemoveContainer" containerID="6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.004378 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} err="failed to get container status \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": rpc error: code = NotFound desc = could not find container \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": container with ID starting with 6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.004398 4810 scope.go:117] "RemoveContainer" containerID="b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.004741 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} err="failed to get container status \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": rpc error: code = NotFound desc = could not find container \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": container with ID starting with b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.004772 4810 scope.go:117] "RemoveContainer" containerID="12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005024 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} err="failed to get container status \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": rpc error: code = NotFound desc = could not find container \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": container with ID starting with 12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005045 4810 scope.go:117] "RemoveContainer" containerID="2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005278 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} err="failed to get container status \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": rpc error: code = NotFound desc = could not find container \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": container with ID starting with 2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005299 4810 scope.go:117] "RemoveContainer" containerID="e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005561 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} err="failed to get container status \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": rpc error: code = NotFound desc = could not find container \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": container with ID starting with e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005578 4810 scope.go:117] "RemoveContainer" containerID="8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005790 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} err="failed to get container status \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": rpc error: code = NotFound desc = could not find container \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": container with ID starting with 8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.005807 4810 scope.go:117] "RemoveContainer" containerID="f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.006091 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412"} err="failed to get container status \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": rpc error: code = NotFound desc = could not find container \"f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412\": container with ID starting with f78096b82fc6ec627a1a9f94b94a5b254374af1936657b18ec82c0321fca5412 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.006141 4810 scope.go:117] "RemoveContainer" containerID="b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.006409 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5"} err="failed to get container status \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": rpc error: code = NotFound desc = could not find container \"b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5\": container with ID starting with b4248ad8c7db005f20afb4ceb75621474e9bb4c50f3584cc5ad4866acb0650e5 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.006427 4810 scope.go:117] "RemoveContainer" containerID="8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.006685 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21"} err="failed to get container status \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": rpc error: code = NotFound desc = could not find container \"8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21\": container with ID starting with 8efbe4e20e34ab535dd03735e893d92dc652006205eee85f42a0488456e89c21 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.006704 4810 scope.go:117] "RemoveContainer" containerID="c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.006950 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546"} err="failed to get container status \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": rpc error: code = NotFound desc = could not find container \"c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546\": container with ID starting with c269fb0789de46ea15d8b9a981faff21e45039b22a837535a4ae1ee0642b8546 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.007057 4810 scope.go:117] "RemoveContainer" containerID="6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.007272 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3"} err="failed to get container status \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": rpc error: code = NotFound desc = could not find container \"6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3\": container with ID starting with 6df17891bb5f3eb357a94f018a472eeb7db828baf19566b876b577b236414ea3 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.007291 4810 scope.go:117] "RemoveContainer" containerID="b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.007542 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4"} err="failed to get container status \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": rpc error: code = NotFound desc = could not find container \"b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4\": container with ID starting with b64be3b5b4655b3e8d8aabf987b19ed0570b586e060c0a3a74e9c3d1a7b113c4 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.007560 4810 scope.go:117] "RemoveContainer" containerID="12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.007861 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45"} err="failed to get container status \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": rpc error: code = NotFound desc = could not find container \"12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45\": container with ID starting with 12da389fd3389842045d22a0c2b9634763526f13fdfaa15c8cddaa04d3e4da45 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.007880 4810 scope.go:117] "RemoveContainer" containerID="2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.008163 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862"} err="failed to get container status \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": rpc error: code = NotFound desc = could not find container \"2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862\": container with ID starting with 2f253a2c27a468d486e50290c71edf3dacc9973ef2da95365767a95f87f66862 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.008181 4810 scope.go:117] "RemoveContainer" containerID="e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.008390 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd"} err="failed to get container status \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": rpc error: code = NotFound desc = could not find container \"e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd\": container with ID starting with e5c9cd8ae60833eb3de3dcd49367f9c070ca266175034877929a6bfbf0ca49dd not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.008425 4810 scope.go:117] "RemoveContainer" containerID="8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.008683 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41"} err="failed to get container status \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": rpc error: code = NotFound desc = could not find container \"8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41\": container with ID starting with 8922e14b82406d9ffb00884a72af0052a6f78b3928f606d820ceeae03e577a41 not found: ID does not exist" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.080108 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee7b17fc-599e-4723-80fb-eb248dfe539d" path="/var/lib/kubelet/pods/ee7b17fc-599e-4723-80fb-eb248dfe539d/volumes" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.649390 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/2.log" Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.653261 4810 generic.go:334] "Generic (PLEG): container finished" podID="5f8702df-efc2-4949-b036-fc8433cc98ea" containerID="14bba5395f8f568a07df7751dda73f81060bfba5cf15be4af64030c64d742cc7" exitCode=0 Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.653303 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerDied","Data":"14bba5395f8f568a07df7751dda73f81060bfba5cf15be4af64030c64d742cc7"} Oct 08 06:41:50 crc kubenswrapper[4810]: I1008 06:41:50.653329 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"2dd2a330ff856e9118a031b1c9e26310211e6430719c9321500b758dc63af867"} Oct 08 06:41:51 crc kubenswrapper[4810]: I1008 06:41:51.661253 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"b79e52cb205bed50b94d8ac78582e62bf5bf5a2195777de3245d116933c948c4"} Oct 08 06:41:51 crc kubenswrapper[4810]: I1008 06:41:51.661610 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"d4445bfa58b8528e95fc30c4466685b375b22472bfbcb5c45b663f7438aadd70"} Oct 08 06:41:51 crc kubenswrapper[4810]: I1008 06:41:51.661624 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"9db47781c69694c1aec8fc56b9fb1c5cdce3139ccbd086b54de026c6397ba040"} Oct 08 06:41:51 crc kubenswrapper[4810]: I1008 06:41:51.661634 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"d548bcc470dce08f684e4becc8b23a0865fff36f51d5176014ad74b6dd5d3679"} Oct 08 06:41:51 crc kubenswrapper[4810]: I1008 06:41:51.661645 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"05d6ecfda185fbf763f0bbcf1e49387c582c1d55466e997cc8083abc8f01416c"} Oct 08 06:41:52 crc kubenswrapper[4810]: I1008 06:41:52.670792 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"6a4c8081eda391d8cd63dc73f80a694dd1fe80501d2ca2b18ad97395eea3294e"} Oct 08 06:41:53 crc kubenswrapper[4810]: I1008 06:41:53.935728 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-2kwbx"] Oct 08 06:41:53 crc kubenswrapper[4810]: I1008 06:41:53.936860 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:53 crc kubenswrapper[4810]: I1008 06:41:53.941300 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 08 06:41:53 crc kubenswrapper[4810]: I1008 06:41:53.941606 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 08 06:41:53 crc kubenswrapper[4810]: I1008 06:41:53.941612 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 08 06:41:53 crc kubenswrapper[4810]: I1008 06:41:53.942543 4810 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-clh98" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.000091 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kqbk\" (UniqueName: \"kubernetes.io/projected/d3673956-faa1-498c-a060-8ec2049f5268-kube-api-access-6kqbk\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.000797 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d3673956-faa1-498c-a060-8ec2049f5268-node-mnt\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.000873 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d3673956-faa1-498c-a060-8ec2049f5268-crc-storage\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.102640 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d3673956-faa1-498c-a060-8ec2049f5268-crc-storage\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.102848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kqbk\" (UniqueName: \"kubernetes.io/projected/d3673956-faa1-498c-a060-8ec2049f5268-kube-api-access-6kqbk\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.102898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d3673956-faa1-498c-a060-8ec2049f5268-node-mnt\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.103408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d3673956-faa1-498c-a060-8ec2049f5268-node-mnt\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.104341 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d3673956-faa1-498c-a060-8ec2049f5268-crc-storage\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.130633 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kqbk\" (UniqueName: \"kubernetes.io/projected/d3673956-faa1-498c-a060-8ec2049f5268-kube-api-access-6kqbk\") pod \"crc-storage-crc-2kwbx\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.264844 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: E1008 06:41:54.304559 4810 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(ab5236f8657357423dbc8e5383079484d770d10656805c1cbeb438fed5dbf405): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 06:41:54 crc kubenswrapper[4810]: E1008 06:41:54.304770 4810 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(ab5236f8657357423dbc8e5383079484d770d10656805c1cbeb438fed5dbf405): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: E1008 06:41:54.304831 4810 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(ab5236f8657357423dbc8e5383079484d770d10656805c1cbeb438fed5dbf405): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:54 crc kubenswrapper[4810]: E1008 06:41:54.304951 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-2kwbx_crc-storage(d3673956-faa1-498c-a060-8ec2049f5268)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-2kwbx_crc-storage(d3673956-faa1-498c-a060-8ec2049f5268)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(ab5236f8657357423dbc8e5383079484d770d10656805c1cbeb438fed5dbf405): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-2kwbx" podUID="d3673956-faa1-498c-a060-8ec2049f5268" Oct 08 06:41:54 crc kubenswrapper[4810]: I1008 06:41:54.690760 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"bf08396a81d7b7f790d31353ec8e8dc91b20f80f0aed18addac59eb18247107f"} Oct 08 06:41:56 crc kubenswrapper[4810]: I1008 06:41:56.708532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" event={"ID":"5f8702df-efc2-4949-b036-fc8433cc98ea","Type":"ContainerStarted","Data":"d6514c2e306b51f01c36bdd4ac8dd0005828ca7d92d04258789268842eb1e443"} Oct 08 06:41:56 crc kubenswrapper[4810]: I1008 06:41:56.709093 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:56 crc kubenswrapper[4810]: I1008 06:41:56.743340 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" podStartSLOduration=7.743306286 podStartE2EDuration="7.743306286s" podCreationTimestamp="2025-10-08 06:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:41:56.74165889 +0000 UTC m=+619.376098640" watchObservedRunningTime="2025-10-08 06:41:56.743306286 +0000 UTC m=+619.377746036" Oct 08 06:41:56 crc kubenswrapper[4810]: I1008 06:41:56.750431 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:57 crc kubenswrapper[4810]: I1008 06:41:57.137389 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2kwbx"] Oct 08 06:41:57 crc kubenswrapper[4810]: I1008 06:41:57.137943 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:57 crc kubenswrapper[4810]: I1008 06:41:57.138558 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:57 crc kubenswrapper[4810]: E1008 06:41:57.180185 4810 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(c5791d712cf74ecaca4a28a7dbcfc9b147669d77797a6b5f3a3c06f1066cad5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 06:41:57 crc kubenswrapper[4810]: E1008 06:41:57.180295 4810 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(c5791d712cf74ecaca4a28a7dbcfc9b147669d77797a6b5f3a3c06f1066cad5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:57 crc kubenswrapper[4810]: E1008 06:41:57.180344 4810 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(c5791d712cf74ecaca4a28a7dbcfc9b147669d77797a6b5f3a3c06f1066cad5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:41:57 crc kubenswrapper[4810]: E1008 06:41:57.180425 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-2kwbx_crc-storage(d3673956-faa1-498c-a060-8ec2049f5268)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-2kwbx_crc-storage(d3673956-faa1-498c-a060-8ec2049f5268)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(c5791d712cf74ecaca4a28a7dbcfc9b147669d77797a6b5f3a3c06f1066cad5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-2kwbx" podUID="d3673956-faa1-498c-a060-8ec2049f5268" Oct 08 06:41:57 crc kubenswrapper[4810]: I1008 06:41:57.718800 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:57 crc kubenswrapper[4810]: I1008 06:41:57.718899 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:41:57 crc kubenswrapper[4810]: I1008 06:41:57.809389 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:42:04 crc kubenswrapper[4810]: I1008 06:42:04.074491 4810 scope.go:117] "RemoveContainer" containerID="97724d3ceaf0c93d52284cdd2d590748da432288355b5517444c117dcb5b06d9" Oct 08 06:42:04 crc kubenswrapper[4810]: E1008 06:42:04.076009 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-txvr6_openshift-multus(9f654c73-c1a2-4292-abe0-830b46ed68f5)\"" pod="openshift-multus/multus-txvr6" podUID="9f654c73-c1a2-4292-abe0-830b46ed68f5" Oct 08 06:42:08 crc kubenswrapper[4810]: I1008 06:42:08.072586 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:08 crc kubenswrapper[4810]: I1008 06:42:08.076248 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:08 crc kubenswrapper[4810]: E1008 06:42:08.114679 4810 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(bdc74b2d47d16ae93e12db4c77b148b904365b65e617ef82af67a493fecc9de1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 06:42:08 crc kubenswrapper[4810]: E1008 06:42:08.115145 4810 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(bdc74b2d47d16ae93e12db4c77b148b904365b65e617ef82af67a493fecc9de1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:08 crc kubenswrapper[4810]: E1008 06:42:08.115181 4810 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(bdc74b2d47d16ae93e12db4c77b148b904365b65e617ef82af67a493fecc9de1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:08 crc kubenswrapper[4810]: E1008 06:42:08.115249 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-2kwbx_crc-storage(d3673956-faa1-498c-a060-8ec2049f5268)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-2kwbx_crc-storage(d3673956-faa1-498c-a060-8ec2049f5268)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2kwbx_crc-storage_d3673956-faa1-498c-a060-8ec2049f5268_0(bdc74b2d47d16ae93e12db4c77b148b904365b65e617ef82af67a493fecc9de1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-2kwbx" podUID="d3673956-faa1-498c-a060-8ec2049f5268" Oct 08 06:42:19 crc kubenswrapper[4810]: I1008 06:42:19.074113 4810 scope.go:117] "RemoveContainer" containerID="97724d3ceaf0c93d52284cdd2d590748da432288355b5517444c117dcb5b06d9" Oct 08 06:42:19 crc kubenswrapper[4810]: I1008 06:42:19.651614 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4b59c" Oct 08 06:42:19 crc kubenswrapper[4810]: I1008 06:42:19.863199 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-txvr6_9f654c73-c1a2-4292-abe0-830b46ed68f5/kube-multus/2.log" Oct 08 06:42:19 crc kubenswrapper[4810]: I1008 06:42:19.863503 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-txvr6" event={"ID":"9f654c73-c1a2-4292-abe0-830b46ed68f5","Type":"ContainerStarted","Data":"31a8819556fd4b10096b2fd324ce3722c28cd69d55b14a90efb90b0c42054565"} Oct 08 06:42:22 crc kubenswrapper[4810]: I1008 06:42:22.073075 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:22 crc kubenswrapper[4810]: I1008 06:42:22.073877 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:22 crc kubenswrapper[4810]: I1008 06:42:22.306580 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2kwbx"] Oct 08 06:42:22 crc kubenswrapper[4810]: W1008 06:42:22.313956 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3673956_faa1_498c_a060_8ec2049f5268.slice/crio-7b088c5403507c44e4599c63a461f2b45f71f2a6c370a9a9963b9b89adaaf7a9 WatchSource:0}: Error finding container 7b088c5403507c44e4599c63a461f2b45f71f2a6c370a9a9963b9b89adaaf7a9: Status 404 returned error can't find the container with id 7b088c5403507c44e4599c63a461f2b45f71f2a6c370a9a9963b9b89adaaf7a9 Oct 08 06:42:22 crc kubenswrapper[4810]: I1008 06:42:22.316220 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 06:42:22 crc kubenswrapper[4810]: I1008 06:42:22.882646 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2kwbx" event={"ID":"d3673956-faa1-498c-a060-8ec2049f5268","Type":"ContainerStarted","Data":"7b088c5403507c44e4599c63a461f2b45f71f2a6c370a9a9963b9b89adaaf7a9"} Oct 08 06:42:24 crc kubenswrapper[4810]: I1008 06:42:24.901397 4810 generic.go:334] "Generic (PLEG): container finished" podID="d3673956-faa1-498c-a060-8ec2049f5268" containerID="61c08f070c761b9efb90c7c9f58685ac2ffc03a3a1a3c2f7be50eb90f84cd3bb" exitCode=0 Oct 08 06:42:24 crc kubenswrapper[4810]: I1008 06:42:24.901544 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2kwbx" event={"ID":"d3673956-faa1-498c-a060-8ec2049f5268","Type":"ContainerDied","Data":"61c08f070c761b9efb90c7c9f58685ac2ffc03a3a1a3c2f7be50eb90f84cd3bb"} Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.235528 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.280313 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kqbk\" (UniqueName: \"kubernetes.io/projected/d3673956-faa1-498c-a060-8ec2049f5268-kube-api-access-6kqbk\") pod \"d3673956-faa1-498c-a060-8ec2049f5268\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.280440 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d3673956-faa1-498c-a060-8ec2049f5268-crc-storage\") pod \"d3673956-faa1-498c-a060-8ec2049f5268\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.280507 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d3673956-faa1-498c-a060-8ec2049f5268-node-mnt\") pod \"d3673956-faa1-498c-a060-8ec2049f5268\" (UID: \"d3673956-faa1-498c-a060-8ec2049f5268\") " Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.280764 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3673956-faa1-498c-a060-8ec2049f5268-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "d3673956-faa1-498c-a060-8ec2049f5268" (UID: "d3673956-faa1-498c-a060-8ec2049f5268"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.287367 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3673956-faa1-498c-a060-8ec2049f5268-kube-api-access-6kqbk" (OuterVolumeSpecName: "kube-api-access-6kqbk") pod "d3673956-faa1-498c-a060-8ec2049f5268" (UID: "d3673956-faa1-498c-a060-8ec2049f5268"). InnerVolumeSpecName "kube-api-access-6kqbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.309088 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3673956-faa1-498c-a060-8ec2049f5268-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "d3673956-faa1-498c-a060-8ec2049f5268" (UID: "d3673956-faa1-498c-a060-8ec2049f5268"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.382329 4810 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/d3673956-faa1-498c-a060-8ec2049f5268-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.382439 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kqbk\" (UniqueName: \"kubernetes.io/projected/d3673956-faa1-498c-a060-8ec2049f5268-kube-api-access-6kqbk\") on node \"crc\" DevicePath \"\"" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.382464 4810 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/d3673956-faa1-498c-a060-8ec2049f5268-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.916523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2kwbx" event={"ID":"d3673956-faa1-498c-a060-8ec2049f5268","Type":"ContainerDied","Data":"7b088c5403507c44e4599c63a461f2b45f71f2a6c370a9a9963b9b89adaaf7a9"} Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.916571 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b088c5403507c44e4599c63a461f2b45f71f2a6c370a9a9963b9b89adaaf7a9" Oct 08 06:42:26 crc kubenswrapper[4810]: I1008 06:42:26.916656 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2kwbx" Oct 08 06:42:32 crc kubenswrapper[4810]: I1008 06:42:32.952196 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w"] Oct 08 06:42:32 crc kubenswrapper[4810]: E1008 06:42:32.952830 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3673956-faa1-498c-a060-8ec2049f5268" containerName="storage" Oct 08 06:42:32 crc kubenswrapper[4810]: I1008 06:42:32.952852 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3673956-faa1-498c-a060-8ec2049f5268" containerName="storage" Oct 08 06:42:32 crc kubenswrapper[4810]: I1008 06:42:32.953040 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3673956-faa1-498c-a060-8ec2049f5268" containerName="storage" Oct 08 06:42:32 crc kubenswrapper[4810]: I1008 06:42:32.954187 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:32 crc kubenswrapper[4810]: I1008 06:42:32.957603 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 06:42:32 crc kubenswrapper[4810]: I1008 06:42:32.969609 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w"] Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.078343 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nj6d\" (UniqueName: \"kubernetes.io/projected/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-kube-api-access-5nj6d\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.078663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.078831 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.179754 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nj6d\" (UniqueName: \"kubernetes.io/projected/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-kube-api-access-5nj6d\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.180155 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.180267 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.180717 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.181216 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.206715 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nj6d\" (UniqueName: \"kubernetes.io/projected/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-kube-api-access-5nj6d\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.277618 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.577847 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w"] Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.956785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" event={"ID":"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c","Type":"ContainerStarted","Data":"94ce69ed756ee873794e6d54bc7604a6fb646596346f31c1759a0eec8078ffff"} Oct 08 06:42:33 crc kubenswrapper[4810]: I1008 06:42:33.957110 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" event={"ID":"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c","Type":"ContainerStarted","Data":"a746ad8a0504298e51f2757ebe1dbabe37fbae077725b20ff5e0bca0f82172a8"} Oct 08 06:42:34 crc kubenswrapper[4810]: I1008 06:42:34.966620 4810 generic.go:334] "Generic (PLEG): container finished" podID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerID="94ce69ed756ee873794e6d54bc7604a6fb646596346f31c1759a0eec8078ffff" exitCode=0 Oct 08 06:42:34 crc kubenswrapper[4810]: I1008 06:42:34.966670 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" event={"ID":"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c","Type":"ContainerDied","Data":"94ce69ed756ee873794e6d54bc7604a6fb646596346f31c1759a0eec8078ffff"} Oct 08 06:42:36 crc kubenswrapper[4810]: I1008 06:42:36.981635 4810 generic.go:334] "Generic (PLEG): container finished" podID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerID="db50a2b839f60281be73dab0365105cf078b63c7834fba52f255745bac3e2e9b" exitCode=0 Oct 08 06:42:36 crc kubenswrapper[4810]: I1008 06:42:36.981693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" event={"ID":"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c","Type":"ContainerDied","Data":"db50a2b839f60281be73dab0365105cf078b63c7834fba52f255745bac3e2e9b"} Oct 08 06:42:37 crc kubenswrapper[4810]: I1008 06:42:37.998796 4810 generic.go:334] "Generic (PLEG): container finished" podID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerID="2c858d7a5e5a08df6998ebf3cec556c30252eb7058832500246af79bff597f99" exitCode=0 Oct 08 06:42:37 crc kubenswrapper[4810]: I1008 06:42:37.998901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" event={"ID":"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c","Type":"ContainerDied","Data":"2c858d7a5e5a08df6998ebf3cec556c30252eb7058832500246af79bff597f99"} Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.314608 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.375956 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-bundle\") pod \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.376452 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nj6d\" (UniqueName: \"kubernetes.io/projected/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-kube-api-access-5nj6d\") pod \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.376563 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-util\") pod \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\" (UID: \"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c\") " Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.377691 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-bundle" (OuterVolumeSpecName: "bundle") pod "93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" (UID: "93a3ddd7-2bd3-4b14-8334-faddbd4cd42c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.383285 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-kube-api-access-5nj6d" (OuterVolumeSpecName: "kube-api-access-5nj6d") pod "93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" (UID: "93a3ddd7-2bd3-4b14-8334-faddbd4cd42c"). InnerVolumeSpecName "kube-api-access-5nj6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.398445 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-util" (OuterVolumeSpecName: "util") pod "93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" (UID: "93a3ddd7-2bd3-4b14-8334-faddbd4cd42c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.477945 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-util\") on node \"crc\" DevicePath \"\"" Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.477998 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:42:39 crc kubenswrapper[4810]: I1008 06:42:39.478014 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nj6d\" (UniqueName: \"kubernetes.io/projected/93a3ddd7-2bd3-4b14-8334-faddbd4cd42c-kube-api-access-5nj6d\") on node \"crc\" DevicePath \"\"" Oct 08 06:42:40 crc kubenswrapper[4810]: I1008 06:42:40.015777 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" event={"ID":"93a3ddd7-2bd3-4b14-8334-faddbd4cd42c","Type":"ContainerDied","Data":"a746ad8a0504298e51f2757ebe1dbabe37fbae077725b20ff5e0bca0f82172a8"} Oct 08 06:42:40 crc kubenswrapper[4810]: I1008 06:42:40.015826 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a746ad8a0504298e51f2757ebe1dbabe37fbae077725b20ff5e0bca0f82172a8" Oct 08 06:42:40 crc kubenswrapper[4810]: I1008 06:42:40.015943 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.743401 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-c75hl"] Oct 08 06:42:41 crc kubenswrapper[4810]: E1008 06:42:41.744088 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerName="pull" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.744106 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerName="pull" Oct 08 06:42:41 crc kubenswrapper[4810]: E1008 06:42:41.744131 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerName="extract" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.744139 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerName="extract" Oct 08 06:42:41 crc kubenswrapper[4810]: E1008 06:42:41.744152 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerName="util" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.744160 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerName="util" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.744286 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a3ddd7-2bd3-4b14-8334-faddbd4cd42c" containerName="extract" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.744816 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.746618 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.746789 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.748291 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jcssp" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.759905 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-c75hl"] Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.811326 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctr5k\" (UniqueName: \"kubernetes.io/projected/65727c84-7a72-4ffb-91bc-93e2242d360a-kube-api-access-ctr5k\") pod \"nmstate-operator-858ddd8f98-c75hl\" (UID: \"65727c84-7a72-4ffb-91bc-93e2242d360a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.912556 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctr5k\" (UniqueName: \"kubernetes.io/projected/65727c84-7a72-4ffb-91bc-93e2242d360a-kube-api-access-ctr5k\") pod \"nmstate-operator-858ddd8f98-c75hl\" (UID: \"65727c84-7a72-4ffb-91bc-93e2242d360a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" Oct 08 06:42:41 crc kubenswrapper[4810]: I1008 06:42:41.936860 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctr5k\" (UniqueName: \"kubernetes.io/projected/65727c84-7a72-4ffb-91bc-93e2242d360a-kube-api-access-ctr5k\") pod \"nmstate-operator-858ddd8f98-c75hl\" (UID: \"65727c84-7a72-4ffb-91bc-93e2242d360a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" Oct 08 06:42:42 crc kubenswrapper[4810]: I1008 06:42:42.063112 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" Oct 08 06:42:42 crc kubenswrapper[4810]: I1008 06:42:42.547063 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-c75hl"] Oct 08 06:42:42 crc kubenswrapper[4810]: W1008 06:42:42.556563 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65727c84_7a72_4ffb_91bc_93e2242d360a.slice/crio-9e35e10a8fd8fb6397931ba8825a5e389e0ee2a2812413ccb15b1ae0d24828af WatchSource:0}: Error finding container 9e35e10a8fd8fb6397931ba8825a5e389e0ee2a2812413ccb15b1ae0d24828af: Status 404 returned error can't find the container with id 9e35e10a8fd8fb6397931ba8825a5e389e0ee2a2812413ccb15b1ae0d24828af Oct 08 06:42:43 crc kubenswrapper[4810]: I1008 06:42:43.038248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" event={"ID":"65727c84-7a72-4ffb-91bc-93e2242d360a","Type":"ContainerStarted","Data":"9e35e10a8fd8fb6397931ba8825a5e389e0ee2a2812413ccb15b1ae0d24828af"} Oct 08 06:42:47 crc kubenswrapper[4810]: I1008 06:42:47.066134 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" event={"ID":"65727c84-7a72-4ffb-91bc-93e2242d360a","Type":"ContainerStarted","Data":"99582fc52385f4259f33273724d85daa593543ac9b22bdd367f7483e9e72e05e"} Oct 08 06:42:47 crc kubenswrapper[4810]: I1008 06:42:47.089273 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-c75hl" podStartSLOduration=1.931579782 podStartE2EDuration="6.089232431s" podCreationTimestamp="2025-10-08 06:42:41 +0000 UTC" firstStartedPulling="2025-10-08 06:42:42.559192456 +0000 UTC m=+665.193632196" lastFinishedPulling="2025-10-08 06:42:46.716845105 +0000 UTC m=+669.351284845" observedRunningTime="2025-10-08 06:42:47.084474602 +0000 UTC m=+669.718914342" watchObservedRunningTime="2025-10-08 06:42:47.089232431 +0000 UTC m=+669.723672201" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.004533 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.005430 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.008149 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-7swnt" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.018591 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.041485 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.042542 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.049908 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.085679 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-rqsl2"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.086836 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.114394 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.127949 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-nmstate-lock\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.128030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zfn2\" (UniqueName: \"kubernetes.io/projected/c1e3d23b-7985-42b6-866b-fd5a93c9921f-kube-api-access-4zfn2\") pod \"nmstate-metrics-fdff9cb8d-z68nd\" (UID: \"c1e3d23b-7985-42b6-866b-fd5a93c9921f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.128078 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-dbus-socket\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.128114 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc9ng\" (UniqueName: \"kubernetes.io/projected/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-kube-api-access-gc9ng\") pod \"nmstate-webhook-6cdbc54649-8x6qm\" (UID: \"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.128141 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg5d8\" (UniqueName: \"kubernetes.io/projected/bfcded37-ea7d-4d11-abc2-d66055730182-kube-api-access-rg5d8\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.128200 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8x6qm\" (UID: \"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.128249 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-ovs-socket\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.188169 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.188975 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.190926 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.191798 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.191866 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-wrk6d" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.199880 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229482 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc9ng\" (UniqueName: \"kubernetes.io/projected/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-kube-api-access-gc9ng\") pod \"nmstate-webhook-6cdbc54649-8x6qm\" (UID: \"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229537 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg5d8\" (UniqueName: \"kubernetes.io/projected/bfcded37-ea7d-4d11-abc2-d66055730182-kube-api-access-rg5d8\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229565 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hrnn\" (UniqueName: \"kubernetes.io/projected/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-kube-api-access-9hrnn\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229598 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8x6qm\" (UID: \"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229625 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229649 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-ovs-socket\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229669 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229708 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-nmstate-lock\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229729 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zfn2\" (UniqueName: \"kubernetes.io/projected/c1e3d23b-7985-42b6-866b-fd5a93c9921f-kube-api-access-4zfn2\") pod \"nmstate-metrics-fdff9cb8d-z68nd\" (UID: \"c1e3d23b-7985-42b6-866b-fd5a93c9921f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.229751 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-dbus-socket\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.230126 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-dbus-socket\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: E1008 06:42:48.230717 4810 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 08 06:42:48 crc kubenswrapper[4810]: E1008 06:42:48.230783 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-tls-key-pair podName:9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba nodeName:}" failed. No retries permitted until 2025-10-08 06:42:48.730762023 +0000 UTC m=+671.365201763 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-tls-key-pair") pod "nmstate-webhook-6cdbc54649-8x6qm" (UID: "9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba") : secret "openshift-nmstate-webhook" not found Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.231003 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-ovs-socket\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.231039 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bfcded37-ea7d-4d11-abc2-d66055730182-nmstate-lock\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.252713 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc9ng\" (UniqueName: \"kubernetes.io/projected/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-kube-api-access-gc9ng\") pod \"nmstate-webhook-6cdbc54649-8x6qm\" (UID: \"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.253211 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg5d8\" (UniqueName: \"kubernetes.io/projected/bfcded37-ea7d-4d11-abc2-d66055730182-kube-api-access-rg5d8\") pod \"nmstate-handler-rqsl2\" (UID: \"bfcded37-ea7d-4d11-abc2-d66055730182\") " pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.253390 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zfn2\" (UniqueName: \"kubernetes.io/projected/c1e3d23b-7985-42b6-866b-fd5a93c9921f-kube-api-access-4zfn2\") pod \"nmstate-metrics-fdff9cb8d-z68nd\" (UID: \"c1e3d23b-7985-42b6-866b-fd5a93c9921f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.330045 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.331586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hrnn\" (UniqueName: \"kubernetes.io/projected/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-kube-api-access-9hrnn\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.331671 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.331700 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: E1008 06:42:48.331842 4810 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 08 06:42:48 crc kubenswrapper[4810]: E1008 06:42:48.331914 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-plugin-serving-cert podName:b2cac821-64bd-4d46-8d3e-e8a51fd9d27f nodeName:}" failed. No retries permitted until 2025-10-08 06:42:48.831894045 +0000 UTC m=+671.466333775 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-62h9n" (UID: "b2cac821-64bd-4d46-8d3e-e8a51fd9d27f") : secret "plugin-serving-cert" not found Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.333095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.364143 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hrnn\" (UniqueName: \"kubernetes.io/projected/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-kube-api-access-9hrnn\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.390425 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-557667c566-fhql4"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.391462 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.416746 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.436022 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-serving-cert\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.436069 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-oauth-config\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.436100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-config\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.436121 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-oauth-serving-cert\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.436189 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8n5b\" (UniqueName: \"kubernetes.io/projected/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-kube-api-access-x8n5b\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.436227 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-trusted-ca-bundle\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.436251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-service-ca\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: W1008 06:42:48.456413 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfcded37_ea7d_4d11_abc2_d66055730182.slice/crio-638169522f855834e078bd221fc7c012105572853a7375a0abdb26709d47d784 WatchSource:0}: Error finding container 638169522f855834e078bd221fc7c012105572853a7375a0abdb26709d47d784: Status 404 returned error can't find the container with id 638169522f855834e078bd221fc7c012105572853a7375a0abdb26709d47d784 Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.462134 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-557667c566-fhql4"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.537241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-trusted-ca-bundle\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.537287 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-service-ca\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.537337 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-serving-cert\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.537353 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-oauth-config\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.537372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-config\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.537388 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-oauth-serving-cert\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.537440 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8n5b\" (UniqueName: \"kubernetes.io/projected/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-kube-api-access-x8n5b\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.538618 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-oauth-serving-cert\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.538673 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-config\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.539360 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-trusted-ca-bundle\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.539597 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-service-ca\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.542189 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-serving-cert\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.546215 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-console-oauth-config\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.559113 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8n5b\" (UniqueName: \"kubernetes.io/projected/c259a3f1-adfe-42cd-8c6b-e2d4904944e6-kube-api-access-x8n5b\") pod \"console-557667c566-fhql4\" (UID: \"c259a3f1-adfe-42cd-8c6b-e2d4904944e6\") " pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.561271 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd"] Oct 08 06:42:48 crc kubenswrapper[4810]: W1008 06:42:48.583447 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1e3d23b_7985_42b6_866b_fd5a93c9921f.slice/crio-09ce5a08c1f81ecf46e03c17a127ea552d3cab8bb92d4ab484c18bfd3ffb9459 WatchSource:0}: Error finding container 09ce5a08c1f81ecf46e03c17a127ea552d3cab8bb92d4ab484c18bfd3ffb9459: Status 404 returned error can't find the container with id 09ce5a08c1f81ecf46e03c17a127ea552d3cab8bb92d4ab484c18bfd3ffb9459 Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.711793 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.741399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8x6qm\" (UID: \"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.746091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8x6qm\" (UID: \"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.842814 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.849453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b2cac821-64bd-4d46-8d3e-e8a51fd9d27f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-62h9n\" (UID: \"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:48 crc kubenswrapper[4810]: W1008 06:42:48.956608 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc259a3f1_adfe_42cd_8c6b_e2d4904944e6.slice/crio-71909fb0561706b3ec54e52bc4ed573fa45297d21fce759c74d37a8001d51ba6 WatchSource:0}: Error finding container 71909fb0561706b3ec54e52bc4ed573fa45297d21fce759c74d37a8001d51ba6: Status 404 returned error can't find the container with id 71909fb0561706b3ec54e52bc4ed573fa45297d21fce759c74d37a8001d51ba6 Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.958544 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-557667c566-fhql4"] Oct 08 06:42:48 crc kubenswrapper[4810]: I1008 06:42:48.960326 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:49 crc kubenswrapper[4810]: I1008 06:42:49.082111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" event={"ID":"c1e3d23b-7985-42b6-866b-fd5a93c9921f","Type":"ContainerStarted","Data":"09ce5a08c1f81ecf46e03c17a127ea552d3cab8bb92d4ab484c18bfd3ffb9459"} Oct 08 06:42:49 crc kubenswrapper[4810]: I1008 06:42:49.082861 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-557667c566-fhql4" event={"ID":"c259a3f1-adfe-42cd-8c6b-e2d4904944e6","Type":"ContainerStarted","Data":"71909fb0561706b3ec54e52bc4ed573fa45297d21fce759c74d37a8001d51ba6"} Oct 08 06:42:49 crc kubenswrapper[4810]: I1008 06:42:49.084070 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rqsl2" event={"ID":"bfcded37-ea7d-4d11-abc2-d66055730182","Type":"ContainerStarted","Data":"638169522f855834e078bd221fc7c012105572853a7375a0abdb26709d47d784"} Oct 08 06:42:49 crc kubenswrapper[4810]: I1008 06:42:49.104422 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" Oct 08 06:42:49 crc kubenswrapper[4810]: I1008 06:42:49.173444 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm"] Oct 08 06:42:49 crc kubenswrapper[4810]: W1008 06:42:49.181696 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d3fe3d3_473d_457f_9e5a_d71de5d1e4ba.slice/crio-763cca52239d9d0f27894fe7d1a4cf0c1604b2f3995e6ce37e5625f012091d27 WatchSource:0}: Error finding container 763cca52239d9d0f27894fe7d1a4cf0c1604b2f3995e6ce37e5625f012091d27: Status 404 returned error can't find the container with id 763cca52239d9d0f27894fe7d1a4cf0c1604b2f3995e6ce37e5625f012091d27 Oct 08 06:42:49 crc kubenswrapper[4810]: I1008 06:42:49.374465 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n"] Oct 08 06:42:50 crc kubenswrapper[4810]: I1008 06:42:50.089688 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" event={"ID":"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f","Type":"ContainerStarted","Data":"5c55e79cb2637a1d46592c8c8aa8455224bf7d6c14bf72aa047bc210f913d894"} Oct 08 06:42:50 crc kubenswrapper[4810]: I1008 06:42:50.090877 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" event={"ID":"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba","Type":"ContainerStarted","Data":"763cca52239d9d0f27894fe7d1a4cf0c1604b2f3995e6ce37e5625f012091d27"} Oct 08 06:42:50 crc kubenswrapper[4810]: I1008 06:42:50.093085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-557667c566-fhql4" event={"ID":"c259a3f1-adfe-42cd-8c6b-e2d4904944e6","Type":"ContainerStarted","Data":"58bc33fc769c2152992af992e795a71cc8376e9b70eea44317d2c7a834b442ee"} Oct 08 06:42:50 crc kubenswrapper[4810]: I1008 06:42:50.120702 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-557667c566-fhql4" podStartSLOduration=2.120676629 podStartE2EDuration="2.120676629s" podCreationTimestamp="2025-10-08 06:42:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:42:50.118520199 +0000 UTC m=+672.752959969" watchObservedRunningTime="2025-10-08 06:42:50.120676629 +0000 UTC m=+672.755116379" Oct 08 06:42:52 crc kubenswrapper[4810]: I1008 06:42:52.108790 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rqsl2" event={"ID":"bfcded37-ea7d-4d11-abc2-d66055730182","Type":"ContainerStarted","Data":"789d52a0901775eaab0b89a798bc164aa1d99e2490696938b84057fdacf34b0e"} Oct 08 06:42:52 crc kubenswrapper[4810]: I1008 06:42:52.109291 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:52 crc kubenswrapper[4810]: I1008 06:42:52.119362 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" event={"ID":"9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba","Type":"ContainerStarted","Data":"4b190356a27e99b0dd90bd9890101449e926298a4b564f78354f81aeac7e5c97"} Oct 08 06:42:52 crc kubenswrapper[4810]: I1008 06:42:52.120486 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:42:52 crc kubenswrapper[4810]: I1008 06:42:52.124458 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" event={"ID":"c1e3d23b-7985-42b6-866b-fd5a93c9921f","Type":"ContainerStarted","Data":"dcdd61962b5ff01edff632a32ee99eef7e9b7e841bf12e9935f7df091582068b"} Oct 08 06:42:52 crc kubenswrapper[4810]: I1008 06:42:52.134157 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-rqsl2" podStartSLOduration=1.5407048319999999 podStartE2EDuration="4.13412918s" podCreationTimestamp="2025-10-08 06:42:48 +0000 UTC" firstStartedPulling="2025-10-08 06:42:48.459397031 +0000 UTC m=+671.093836771" lastFinishedPulling="2025-10-08 06:42:51.052821369 +0000 UTC m=+673.687261119" observedRunningTime="2025-10-08 06:42:52.127673453 +0000 UTC m=+674.762113223" watchObservedRunningTime="2025-10-08 06:42:52.13412918 +0000 UTC m=+674.768568920" Oct 08 06:42:52 crc kubenswrapper[4810]: I1008 06:42:52.160274 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" podStartSLOduration=2.293319091 podStartE2EDuration="4.160244376s" podCreationTimestamp="2025-10-08 06:42:48 +0000 UTC" firstStartedPulling="2025-10-08 06:42:49.187624872 +0000 UTC m=+671.822064612" lastFinishedPulling="2025-10-08 06:42:51.054550147 +0000 UTC m=+673.688989897" observedRunningTime="2025-10-08 06:42:52.15238005 +0000 UTC m=+674.786819820" watchObservedRunningTime="2025-10-08 06:42:52.160244376 +0000 UTC m=+674.794684136" Oct 08 06:42:53 crc kubenswrapper[4810]: I1008 06:42:53.147059 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" event={"ID":"b2cac821-64bd-4d46-8d3e-e8a51fd9d27f","Type":"ContainerStarted","Data":"d88c209bf937e0051278d9873553ad99d24e86fcc10baec53a901793ed5e38a5"} Oct 08 06:42:53 crc kubenswrapper[4810]: I1008 06:42:53.167600 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-62h9n" podStartSLOduration=2.387400349 podStartE2EDuration="5.167560958s" podCreationTimestamp="2025-10-08 06:42:48 +0000 UTC" firstStartedPulling="2025-10-08 06:42:49.383228153 +0000 UTC m=+672.017667893" lastFinishedPulling="2025-10-08 06:42:52.163388762 +0000 UTC m=+674.797828502" observedRunningTime="2025-10-08 06:42:53.165759179 +0000 UTC m=+675.800198919" watchObservedRunningTime="2025-10-08 06:42:53.167560958 +0000 UTC m=+675.802000698" Oct 08 06:42:54 crc kubenswrapper[4810]: I1008 06:42:54.156730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" event={"ID":"c1e3d23b-7985-42b6-866b-fd5a93c9921f","Type":"ContainerStarted","Data":"e3d9ef9a248de47d13b519bff67bf71bfeb5bbc8578779232098207a8d65a331"} Oct 08 06:42:54 crc kubenswrapper[4810]: I1008 06:42:54.187283 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-z68nd" podStartSLOduration=2.290711927 podStartE2EDuration="7.187254279s" podCreationTimestamp="2025-10-08 06:42:47 +0000 UTC" firstStartedPulling="2025-10-08 06:42:48.586288238 +0000 UTC m=+671.220727978" lastFinishedPulling="2025-10-08 06:42:53.48283059 +0000 UTC m=+676.117270330" observedRunningTime="2025-10-08 06:42:54.186399116 +0000 UTC m=+676.820838856" watchObservedRunningTime="2025-10-08 06:42:54.187254279 +0000 UTC m=+676.821694069" Oct 08 06:42:58 crc kubenswrapper[4810]: I1008 06:42:58.459089 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-rqsl2" Oct 08 06:42:58 crc kubenswrapper[4810]: I1008 06:42:58.714125 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:58 crc kubenswrapper[4810]: I1008 06:42:58.714256 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:58 crc kubenswrapper[4810]: I1008 06:42:58.722733 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:59 crc kubenswrapper[4810]: I1008 06:42:59.202000 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-557667c566-fhql4" Oct 08 06:42:59 crc kubenswrapper[4810]: I1008 06:42:59.325225 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fq7wx"] Oct 08 06:43:08 crc kubenswrapper[4810]: I1008 06:43:08.969137 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8x6qm" Oct 08 06:43:19 crc kubenswrapper[4810]: I1008 06:43:19.431874 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:43:19 crc kubenswrapper[4810]: I1008 06:43:19.433105 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.371883 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-fq7wx" podUID="303e572d-0bdf-4cf2-85bf-976e5ac59130" containerName="console" containerID="cri-o://568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d" gracePeriod=15 Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.770561 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fq7wx_303e572d-0bdf-4cf2-85bf-976e5ac59130/console/0.log" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.770867 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.837668 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-serving-cert\") pod \"303e572d-0bdf-4cf2-85bf-976e5ac59130\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.837724 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-trusted-ca-bundle\") pod \"303e572d-0bdf-4cf2-85bf-976e5ac59130\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.837752 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-config\") pod \"303e572d-0bdf-4cf2-85bf-976e5ac59130\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.837787 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-oauth-config\") pod \"303e572d-0bdf-4cf2-85bf-976e5ac59130\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.837838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-oauth-serving-cert\") pod \"303e572d-0bdf-4cf2-85bf-976e5ac59130\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.837917 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-service-ca\") pod \"303e572d-0bdf-4cf2-85bf-976e5ac59130\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.837936 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8457\" (UniqueName: \"kubernetes.io/projected/303e572d-0bdf-4cf2-85bf-976e5ac59130-kube-api-access-g8457\") pod \"303e572d-0bdf-4cf2-85bf-976e5ac59130\" (UID: \"303e572d-0bdf-4cf2-85bf-976e5ac59130\") " Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.839119 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-service-ca" (OuterVolumeSpecName: "service-ca") pod "303e572d-0bdf-4cf2-85bf-976e5ac59130" (UID: "303e572d-0bdf-4cf2-85bf-976e5ac59130"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.839133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "303e572d-0bdf-4cf2-85bf-976e5ac59130" (UID: "303e572d-0bdf-4cf2-85bf-976e5ac59130"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.839485 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "303e572d-0bdf-4cf2-85bf-976e5ac59130" (UID: "303e572d-0bdf-4cf2-85bf-976e5ac59130"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.841636 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-config" (OuterVolumeSpecName: "console-config") pod "303e572d-0bdf-4cf2-85bf-976e5ac59130" (UID: "303e572d-0bdf-4cf2-85bf-976e5ac59130"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.855082 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/303e572d-0bdf-4cf2-85bf-976e5ac59130-kube-api-access-g8457" (OuterVolumeSpecName: "kube-api-access-g8457") pod "303e572d-0bdf-4cf2-85bf-976e5ac59130" (UID: "303e572d-0bdf-4cf2-85bf-976e5ac59130"). InnerVolumeSpecName "kube-api-access-g8457". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.862008 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "303e572d-0bdf-4cf2-85bf-976e5ac59130" (UID: "303e572d-0bdf-4cf2-85bf-976e5ac59130"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.863213 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "303e572d-0bdf-4cf2-85bf-976e5ac59130" (UID: "303e572d-0bdf-4cf2-85bf-976e5ac59130"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.939956 4810 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.940013 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.940024 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8457\" (UniqueName: \"kubernetes.io/projected/303e572d-0bdf-4cf2-85bf-976e5ac59130-kube-api-access-g8457\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.940037 4810 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.940050 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.940061 4810 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:24 crc kubenswrapper[4810]: I1008 06:43:24.940073 4810 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/303e572d-0bdf-4cf2-85bf-976e5ac59130-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.184909 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk"] Oct 08 06:43:25 crc kubenswrapper[4810]: E1008 06:43:25.185787 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303e572d-0bdf-4cf2-85bf-976e5ac59130" containerName="console" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.185812 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="303e572d-0bdf-4cf2-85bf-976e5ac59130" containerName="console" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.186079 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="303e572d-0bdf-4cf2-85bf-976e5ac59130" containerName="console" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.187670 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.190731 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.194588 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk"] Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.244755 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.244852 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swn28\" (UniqueName: \"kubernetes.io/projected/a378b60b-7299-43c4-b159-5b5c54e41b11-kube-api-access-swn28\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.244891 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.346138 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.346217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swn28\" (UniqueName: \"kubernetes.io/projected/a378b60b-7299-43c4-b159-5b5c54e41b11-kube-api-access-swn28\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.346246 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.346841 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.347222 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.364312 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swn28\" (UniqueName: \"kubernetes.io/projected/a378b60b-7299-43c4-b159-5b5c54e41b11-kube-api-access-swn28\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.402665 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fq7wx_303e572d-0bdf-4cf2-85bf-976e5ac59130/console/0.log" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.402736 4810 generic.go:334] "Generic (PLEG): container finished" podID="303e572d-0bdf-4cf2-85bf-976e5ac59130" containerID="568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d" exitCode=2 Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.402774 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fq7wx" event={"ID":"303e572d-0bdf-4cf2-85bf-976e5ac59130","Type":"ContainerDied","Data":"568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d"} Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.402815 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fq7wx" event={"ID":"303e572d-0bdf-4cf2-85bf-976e5ac59130","Type":"ContainerDied","Data":"77a7dcf6b9070130148865aa937ace541edcdd74ad054460aecc7de24ec4ffaa"} Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.402834 4810 scope.go:117] "RemoveContainer" containerID="568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.402871 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fq7wx" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.436264 4810 scope.go:117] "RemoveContainer" containerID="568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d" Oct 08 06:43:25 crc kubenswrapper[4810]: E1008 06:43:25.437141 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d\": container with ID starting with 568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d not found: ID does not exist" containerID="568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.437290 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d"} err="failed to get container status \"568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d\": rpc error: code = NotFound desc = could not find container \"568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d\": container with ID starting with 568a85565c5ba985549d9c52cd9daa33763994b62ece267be42d4de2bf21189d not found: ID does not exist" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.450360 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fq7wx"] Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.454729 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-fq7wx"] Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.512829 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:25 crc kubenswrapper[4810]: I1008 06:43:25.702342 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk"] Oct 08 06:43:26 crc kubenswrapper[4810]: I1008 06:43:26.080100 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="303e572d-0bdf-4cf2-85bf-976e5ac59130" path="/var/lib/kubelet/pods/303e572d-0bdf-4cf2-85bf-976e5ac59130/volumes" Oct 08 06:43:26 crc kubenswrapper[4810]: I1008 06:43:26.414475 4810 generic.go:334] "Generic (PLEG): container finished" podID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerID="b27c3cde95bd7c20699b706bff691814c6eeaeb90b7fd579c53a690677324722" exitCode=0 Oct 08 06:43:26 crc kubenswrapper[4810]: I1008 06:43:26.414533 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" event={"ID":"a378b60b-7299-43c4-b159-5b5c54e41b11","Type":"ContainerDied","Data":"b27c3cde95bd7c20699b706bff691814c6eeaeb90b7fd579c53a690677324722"} Oct 08 06:43:26 crc kubenswrapper[4810]: I1008 06:43:26.414571 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" event={"ID":"a378b60b-7299-43c4-b159-5b5c54e41b11","Type":"ContainerStarted","Data":"5f0906117217ee64813e81abc16a59fed242a51b8053d19dc02312c2f724713b"} Oct 08 06:43:28 crc kubenswrapper[4810]: I1008 06:43:28.431499 4810 generic.go:334] "Generic (PLEG): container finished" podID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerID="278af774d3b4153a43105715ac469e9c9d88cc61b624808fb6670cd6f3ac2387" exitCode=0 Oct 08 06:43:28 crc kubenswrapper[4810]: I1008 06:43:28.431599 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" event={"ID":"a378b60b-7299-43c4-b159-5b5c54e41b11","Type":"ContainerDied","Data":"278af774d3b4153a43105715ac469e9c9d88cc61b624808fb6670cd6f3ac2387"} Oct 08 06:43:29 crc kubenswrapper[4810]: I1008 06:43:29.443557 4810 generic.go:334] "Generic (PLEG): container finished" podID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerID="048215982f3fe9215c835a7abea829e9647a2266b82d179d64ce3014f4019941" exitCode=0 Oct 08 06:43:29 crc kubenswrapper[4810]: I1008 06:43:29.443671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" event={"ID":"a378b60b-7299-43c4-b159-5b5c54e41b11","Type":"ContainerDied","Data":"048215982f3fe9215c835a7abea829e9647a2266b82d179d64ce3014f4019941"} Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.731523 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.832117 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-bundle\") pod \"a378b60b-7299-43c4-b159-5b5c54e41b11\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.832224 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-util\") pod \"a378b60b-7299-43c4-b159-5b5c54e41b11\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.832467 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swn28\" (UniqueName: \"kubernetes.io/projected/a378b60b-7299-43c4-b159-5b5c54e41b11-kube-api-access-swn28\") pod \"a378b60b-7299-43c4-b159-5b5c54e41b11\" (UID: \"a378b60b-7299-43c4-b159-5b5c54e41b11\") " Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.833473 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-bundle" (OuterVolumeSpecName: "bundle") pod "a378b60b-7299-43c4-b159-5b5c54e41b11" (UID: "a378b60b-7299-43c4-b159-5b5c54e41b11"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.843502 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a378b60b-7299-43c4-b159-5b5c54e41b11-kube-api-access-swn28" (OuterVolumeSpecName: "kube-api-access-swn28") pod "a378b60b-7299-43c4-b159-5b5c54e41b11" (UID: "a378b60b-7299-43c4-b159-5b5c54e41b11"). InnerVolumeSpecName "kube-api-access-swn28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.852696 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-util" (OuterVolumeSpecName: "util") pod "a378b60b-7299-43c4-b159-5b5c54e41b11" (UID: "a378b60b-7299-43c4-b159-5b5c54e41b11"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.934570 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swn28\" (UniqueName: \"kubernetes.io/projected/a378b60b-7299-43c4-b159-5b5c54e41b11-kube-api-access-swn28\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.934642 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:30 crc kubenswrapper[4810]: I1008 06:43:30.934668 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a378b60b-7299-43c4-b159-5b5c54e41b11-util\") on node \"crc\" DevicePath \"\"" Oct 08 06:43:31 crc kubenswrapper[4810]: I1008 06:43:31.462858 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" event={"ID":"a378b60b-7299-43c4-b159-5b5c54e41b11","Type":"ContainerDied","Data":"5f0906117217ee64813e81abc16a59fed242a51b8053d19dc02312c2f724713b"} Oct 08 06:43:31 crc kubenswrapper[4810]: I1008 06:43:31.462907 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f0906117217ee64813e81abc16a59fed242a51b8053d19dc02312c2f724713b" Oct 08 06:43:31 crc kubenswrapper[4810]: I1008 06:43:31.463014 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.197132 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db"] Oct 08 06:43:40 crc kubenswrapper[4810]: E1008 06:43:40.197904 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerName="extract" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.197917 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerName="extract" Oct 08 06:43:40 crc kubenswrapper[4810]: E1008 06:43:40.197925 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerName="util" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.197931 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerName="util" Oct 08 06:43:40 crc kubenswrapper[4810]: E1008 06:43:40.197948 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerName="pull" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.197953 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerName="pull" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.198083 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a378b60b-7299-43c4-b159-5b5c54e41b11" containerName="extract" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.198508 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.202840 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.203200 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.206493 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.206532 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.206918 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-jqsqj" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.235065 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db"] Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.269624 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1cf28dd8-f466-44e0-b814-b956dda73d9c-apiservice-cert\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.269692 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8thd\" (UniqueName: \"kubernetes.io/projected/1cf28dd8-f466-44e0-b814-b956dda73d9c-kube-api-access-v8thd\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.269711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1cf28dd8-f466-44e0-b814-b956dda73d9c-webhook-cert\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.370497 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1cf28dd8-f466-44e0-b814-b956dda73d9c-apiservice-cert\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.370613 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1cf28dd8-f466-44e0-b814-b956dda73d9c-webhook-cert\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.370633 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8thd\" (UniqueName: \"kubernetes.io/projected/1cf28dd8-f466-44e0-b814-b956dda73d9c-kube-api-access-v8thd\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.379064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1cf28dd8-f466-44e0-b814-b956dda73d9c-webhook-cert\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.379079 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1cf28dd8-f466-44e0-b814-b956dda73d9c-apiservice-cert\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.391246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8thd\" (UniqueName: \"kubernetes.io/projected/1cf28dd8-f466-44e0-b814-b956dda73d9c-kube-api-access-v8thd\") pod \"metallb-operator-controller-manager-85f4bbfddc-bx2db\" (UID: \"1cf28dd8-f466-44e0-b814-b956dda73d9c\") " pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.516477 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.518207 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh"] Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.518906 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.520932 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.521382 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-4zv25" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.521943 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.533448 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh"] Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.573034 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-apiservice-cert\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.573091 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvtxs\" (UniqueName: \"kubernetes.io/projected/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-kube-api-access-hvtxs\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.573119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-webhook-cert\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.676398 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-apiservice-cert\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.676875 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvtxs\" (UniqueName: \"kubernetes.io/projected/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-kube-api-access-hvtxs\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.676908 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-webhook-cert\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.680437 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-webhook-cert\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.694798 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-apiservice-cert\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.698299 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvtxs\" (UniqueName: \"kubernetes.io/projected/08e95f35-af58-4c5a-9d5c-3fb6dd406a55-kube-api-access-hvtxs\") pod \"metallb-operator-webhook-server-75dc598b55-b4qjh\" (UID: \"08e95f35-af58-4c5a-9d5c-3fb6dd406a55\") " pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.783452 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db"] Oct 08 06:43:40 crc kubenswrapper[4810]: W1008 06:43:40.786450 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cf28dd8_f466_44e0_b814_b956dda73d9c.slice/crio-63b29a248984091f7a7f6da7d8a55ece2460631da056799fe1bd3558bb33a4cc WatchSource:0}: Error finding container 63b29a248984091f7a7f6da7d8a55ece2460631da056799fe1bd3558bb33a4cc: Status 404 returned error can't find the container with id 63b29a248984091f7a7f6da7d8a55ece2460631da056799fe1bd3558bb33a4cc Oct 08 06:43:40 crc kubenswrapper[4810]: I1008 06:43:40.837201 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:41 crc kubenswrapper[4810]: I1008 06:43:41.119767 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh"] Oct 08 06:43:41 crc kubenswrapper[4810]: W1008 06:43:41.124361 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08e95f35_af58_4c5a_9d5c_3fb6dd406a55.slice/crio-2b385a76d70d76be86b4127b2b99557f53f2e174b2dba0476c839036ee6f6784 WatchSource:0}: Error finding container 2b385a76d70d76be86b4127b2b99557f53f2e174b2dba0476c839036ee6f6784: Status 404 returned error can't find the container with id 2b385a76d70d76be86b4127b2b99557f53f2e174b2dba0476c839036ee6f6784 Oct 08 06:43:41 crc kubenswrapper[4810]: I1008 06:43:41.521699 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" event={"ID":"1cf28dd8-f466-44e0-b814-b956dda73d9c","Type":"ContainerStarted","Data":"63b29a248984091f7a7f6da7d8a55ece2460631da056799fe1bd3558bb33a4cc"} Oct 08 06:43:41 crc kubenswrapper[4810]: I1008 06:43:41.522833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" event={"ID":"08e95f35-af58-4c5a-9d5c-3fb6dd406a55","Type":"ContainerStarted","Data":"2b385a76d70d76be86b4127b2b99557f53f2e174b2dba0476c839036ee6f6784"} Oct 08 06:43:46 crc kubenswrapper[4810]: I1008 06:43:46.557676 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" event={"ID":"1cf28dd8-f466-44e0-b814-b956dda73d9c","Type":"ContainerStarted","Data":"76d34b8186e0bcf70d042240bce44e77709a48d0a21154a7fdb858259750bf9c"} Oct 08 06:43:46 crc kubenswrapper[4810]: I1008 06:43:46.558267 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:43:46 crc kubenswrapper[4810]: I1008 06:43:46.560362 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" event={"ID":"08e95f35-af58-4c5a-9d5c-3fb6dd406a55","Type":"ContainerStarted","Data":"a214b676dcfa32df820947b45150de403d7d837ad43311a961f6b4be764f2100"} Oct 08 06:43:46 crc kubenswrapper[4810]: I1008 06:43:46.560525 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:43:46 crc kubenswrapper[4810]: I1008 06:43:46.626547 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" podStartSLOduration=1.77493752 podStartE2EDuration="6.626530149s" podCreationTimestamp="2025-10-08 06:43:40 +0000 UTC" firstStartedPulling="2025-10-08 06:43:40.789391215 +0000 UTC m=+723.423830955" lastFinishedPulling="2025-10-08 06:43:45.640983844 +0000 UTC m=+728.275423584" observedRunningTime="2025-10-08 06:43:46.59191809 +0000 UTC m=+729.226357820" watchObservedRunningTime="2025-10-08 06:43:46.626530149 +0000 UTC m=+729.260969889" Oct 08 06:43:46 crc kubenswrapper[4810]: I1008 06:43:46.627434 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" podStartSLOduration=2.110921479 podStartE2EDuration="6.627430014s" podCreationTimestamp="2025-10-08 06:43:40 +0000 UTC" firstStartedPulling="2025-10-08 06:43:41.127952835 +0000 UTC m=+723.762392575" lastFinishedPulling="2025-10-08 06:43:45.64446137 +0000 UTC m=+728.278901110" observedRunningTime="2025-10-08 06:43:46.623803684 +0000 UTC m=+729.258243424" watchObservedRunningTime="2025-10-08 06:43:46.627430014 +0000 UTC m=+729.261869754" Oct 08 06:43:49 crc kubenswrapper[4810]: I1008 06:43:49.431731 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:43:49 crc kubenswrapper[4810]: I1008 06:43:49.432300 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:44:00 crc kubenswrapper[4810]: I1008 06:44:00.845041 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-75dc598b55-b4qjh" Oct 08 06:44:03 crc kubenswrapper[4810]: I1008 06:44:03.901277 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-56cpr"] Oct 08 06:44:03 crc kubenswrapper[4810]: I1008 06:44:03.903519 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" podUID="4980c5d8-c614-4943-aff6-cb07dac7593f" containerName="controller-manager" containerID="cri-o://ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd" gracePeriod=30 Oct 08 06:44:03 crc kubenswrapper[4810]: I1008 06:44:03.975806 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s"] Oct 08 06:44:03 crc kubenswrapper[4810]: I1008 06:44:03.976175 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" podUID="9b63477b-8171-4222-815b-16c7b5ba0e13" containerName="route-controller-manager" containerID="cri-o://f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae" gracePeriod=30 Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.464732 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.471394 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.564930 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-config\") pod \"9b63477b-8171-4222-815b-16c7b5ba0e13\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565016 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4980c5d8-c614-4943-aff6-cb07dac7593f-serving-cert\") pod \"4980c5d8-c614-4943-aff6-cb07dac7593f\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565076 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-config\") pod \"4980c5d8-c614-4943-aff6-cb07dac7593f\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565114 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-proxy-ca-bundles\") pod \"4980c5d8-c614-4943-aff6-cb07dac7593f\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565132 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-client-ca\") pod \"4980c5d8-c614-4943-aff6-cb07dac7593f\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565159 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-client-ca\") pod \"9b63477b-8171-4222-815b-16c7b5ba0e13\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565186 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b63477b-8171-4222-815b-16c7b5ba0e13-serving-cert\") pod \"9b63477b-8171-4222-815b-16c7b5ba0e13\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565208 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzpz6\" (UniqueName: \"kubernetes.io/projected/4980c5d8-c614-4943-aff6-cb07dac7593f-kube-api-access-rzpz6\") pod \"4980c5d8-c614-4943-aff6-cb07dac7593f\" (UID: \"4980c5d8-c614-4943-aff6-cb07dac7593f\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.565256 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88xlc\" (UniqueName: \"kubernetes.io/projected/9b63477b-8171-4222-815b-16c7b5ba0e13-kube-api-access-88xlc\") pod \"9b63477b-8171-4222-815b-16c7b5ba0e13\" (UID: \"9b63477b-8171-4222-815b-16c7b5ba0e13\") " Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.566840 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-client-ca" (OuterVolumeSpecName: "client-ca") pod "9b63477b-8171-4222-815b-16c7b5ba0e13" (UID: "9b63477b-8171-4222-815b-16c7b5ba0e13"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.566900 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-client-ca" (OuterVolumeSpecName: "client-ca") pod "4980c5d8-c614-4943-aff6-cb07dac7593f" (UID: "4980c5d8-c614-4943-aff6-cb07dac7593f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.567300 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-config" (OuterVolumeSpecName: "config") pod "9b63477b-8171-4222-815b-16c7b5ba0e13" (UID: "9b63477b-8171-4222-815b-16c7b5ba0e13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.567429 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-config" (OuterVolumeSpecName: "config") pod "4980c5d8-c614-4943-aff6-cb07dac7593f" (UID: "4980c5d8-c614-4943-aff6-cb07dac7593f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.567458 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4980c5d8-c614-4943-aff6-cb07dac7593f" (UID: "4980c5d8-c614-4943-aff6-cb07dac7593f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.574313 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4980c5d8-c614-4943-aff6-cb07dac7593f-kube-api-access-rzpz6" (OuterVolumeSpecName: "kube-api-access-rzpz6") pod "4980c5d8-c614-4943-aff6-cb07dac7593f" (UID: "4980c5d8-c614-4943-aff6-cb07dac7593f"). InnerVolumeSpecName "kube-api-access-rzpz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.575564 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4980c5d8-c614-4943-aff6-cb07dac7593f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4980c5d8-c614-4943-aff6-cb07dac7593f" (UID: "4980c5d8-c614-4943-aff6-cb07dac7593f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.577386 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b63477b-8171-4222-815b-16c7b5ba0e13-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9b63477b-8171-4222-815b-16c7b5ba0e13" (UID: "9b63477b-8171-4222-815b-16c7b5ba0e13"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.577554 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b63477b-8171-4222-815b-16c7b5ba0e13-kube-api-access-88xlc" (OuterVolumeSpecName: "kube-api-access-88xlc") pod "9b63477b-8171-4222-815b-16c7b5ba0e13" (UID: "9b63477b-8171-4222-815b-16c7b5ba0e13"). InnerVolumeSpecName "kube-api-access-88xlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666596 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88xlc\" (UniqueName: \"kubernetes.io/projected/9b63477b-8171-4222-815b-16c7b5ba0e13-kube-api-access-88xlc\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666640 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666650 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4980c5d8-c614-4943-aff6-cb07dac7593f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666661 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666670 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666678 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4980c5d8-c614-4943-aff6-cb07dac7593f-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666685 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b63477b-8171-4222-815b-16c7b5ba0e13-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666693 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b63477b-8171-4222-815b-16c7b5ba0e13-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.666701 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzpz6\" (UniqueName: \"kubernetes.io/projected/4980c5d8-c614-4943-aff6-cb07dac7593f-kube-api-access-rzpz6\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.685453 4810 generic.go:334] "Generic (PLEG): container finished" podID="9b63477b-8171-4222-815b-16c7b5ba0e13" containerID="f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae" exitCode=0 Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.685525 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.685578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" event={"ID":"9b63477b-8171-4222-815b-16c7b5ba0e13","Type":"ContainerDied","Data":"f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae"} Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.685660 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s" event={"ID":"9b63477b-8171-4222-815b-16c7b5ba0e13","Type":"ContainerDied","Data":"d2d94e59c1d8845e19085dcd4b276596f9aebbf17434cfd8ed8a8d3013124f8a"} Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.685698 4810 scope.go:117] "RemoveContainer" containerID="f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.689401 4810 generic.go:334] "Generic (PLEG): container finished" podID="4980c5d8-c614-4943-aff6-cb07dac7593f" containerID="ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd" exitCode=0 Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.689523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" event={"ID":"4980c5d8-c614-4943-aff6-cb07dac7593f","Type":"ContainerDied","Data":"ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd"} Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.689642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" event={"ID":"4980c5d8-c614-4943-aff6-cb07dac7593f","Type":"ContainerDied","Data":"cc96bd6050602c87e392a5b2376b63cee164e01ef72abb779421d9cbbfa33f02"} Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.689786 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-56cpr" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.706483 4810 scope.go:117] "RemoveContainer" containerID="f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae" Oct 08 06:44:04 crc kubenswrapper[4810]: E1008 06:44:04.707850 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae\": container with ID starting with f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae not found: ID does not exist" containerID="f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.707916 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae"} err="failed to get container status \"f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae\": rpc error: code = NotFound desc = could not find container \"f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae\": container with ID starting with f7e57cd6d891b2f8651a1cbd05409bc31a76f2a3598a3fb2f2a998e4f9b3c8ae not found: ID does not exist" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.707947 4810 scope.go:117] "RemoveContainer" containerID="ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.720614 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s"] Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.728555 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7zw2s"] Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.729029 4810 scope.go:117] "RemoveContainer" containerID="ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.736876 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-56cpr"] Oct 08 06:44:04 crc kubenswrapper[4810]: E1008 06:44:04.738286 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd\": container with ID starting with ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd not found: ID does not exist" containerID="ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.738340 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd"} err="failed to get container status \"ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd\": rpc error: code = NotFound desc = could not find container \"ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd\": container with ID starting with ad46d74b8155782eeec201f65d6049db526fa9a7ad4886669d870aec88d708dd not found: ID does not exist" Oct 08 06:44:04 crc kubenswrapper[4810]: I1008 06:44:04.748427 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-56cpr"] Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.485377 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm"] Oct 08 06:44:05 crc kubenswrapper[4810]: E1008 06:44:05.485595 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b63477b-8171-4222-815b-16c7b5ba0e13" containerName="route-controller-manager" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.485608 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b63477b-8171-4222-815b-16c7b5ba0e13" containerName="route-controller-manager" Oct 08 06:44:05 crc kubenswrapper[4810]: E1008 06:44:05.485615 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4980c5d8-c614-4943-aff6-cb07dac7593f" containerName="controller-manager" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.485621 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4980c5d8-c614-4943-aff6-cb07dac7593f" containerName="controller-manager" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.485700 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4980c5d8-c614-4943-aff6-cb07dac7593f" containerName="controller-manager" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.485713 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b63477b-8171-4222-815b-16c7b5ba0e13" containerName="route-controller-manager" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.486075 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.490554 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.490893 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.491055 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.491596 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.492354 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.494395 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr"] Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.495029 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.500480 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.500535 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.500554 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.500694 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.501142 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.501392 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.503600 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.506182 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm"] Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.507907 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.522587 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr"] Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.577945 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bceec37-623a-4649-b751-d2ce4db6d764-client-ca\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff29c930-8476-47c0-b258-e7756f0e28fb-serving-cert\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578045 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bceec37-623a-4649-b751-d2ce4db6d764-serving-cert\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578065 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-client-ca\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578084 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-proxy-ca-bundles\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578106 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bceec37-623a-4649-b751-d2ce4db6d764-config\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578121 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9rg2\" (UniqueName: \"kubernetes.io/projected/ff29c930-8476-47c0-b258-e7756f0e28fb-kube-api-access-c9rg2\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578317 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-config\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.578459 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw2z7\" (UniqueName: \"kubernetes.io/projected/3bceec37-623a-4649-b751-d2ce4db6d764-kube-api-access-jw2z7\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bceec37-623a-4649-b751-d2ce4db6d764-client-ca\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680360 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff29c930-8476-47c0-b258-e7756f0e28fb-serving-cert\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bceec37-623a-4649-b751-d2ce4db6d764-serving-cert\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680408 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-client-ca\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680423 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-proxy-ca-bundles\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680447 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bceec37-623a-4649-b751-d2ce4db6d764-config\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9rg2\" (UniqueName: \"kubernetes.io/projected/ff29c930-8476-47c0-b258-e7756f0e28fb-kube-api-access-c9rg2\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680493 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-config\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.680523 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw2z7\" (UniqueName: \"kubernetes.io/projected/3bceec37-623a-4649-b751-d2ce4db6d764-kube-api-access-jw2z7\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.681959 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bceec37-623a-4649-b751-d2ce4db6d764-config\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.682044 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-proxy-ca-bundles\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.682044 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-client-ca\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.683119 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff29c930-8476-47c0-b258-e7756f0e28fb-config\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.683786 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3bceec37-623a-4649-b751-d2ce4db6d764-client-ca\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.685802 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff29c930-8476-47c0-b258-e7756f0e28fb-serving-cert\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.687452 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bceec37-623a-4649-b751-d2ce4db6d764-serving-cert\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.699445 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw2z7\" (UniqueName: \"kubernetes.io/projected/3bceec37-623a-4649-b751-d2ce4db6d764-kube-api-access-jw2z7\") pod \"route-controller-manager-5f7b67b7bd-jk4dr\" (UID: \"3bceec37-623a-4649-b751-d2ce4db6d764\") " pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.705208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9rg2\" (UniqueName: \"kubernetes.io/projected/ff29c930-8476-47c0-b258-e7756f0e28fb-kube-api-access-c9rg2\") pod \"controller-manager-69fdb48b8b-hkxgm\" (UID: \"ff29c930-8476-47c0-b258-e7756f0e28fb\") " pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.802847 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:05 crc kubenswrapper[4810]: I1008 06:44:05.811644 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.057886 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr"] Oct 08 06:44:06 crc kubenswrapper[4810]: W1008 06:44:06.094354 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bceec37_623a_4649_b751_d2ce4db6d764.slice/crio-11138c0ccf75f4e3f44c04a831a47ce0eab6ea00a1426987234c898575d081f6 WatchSource:0}: Error finding container 11138c0ccf75f4e3f44c04a831a47ce0eab6ea00a1426987234c898575d081f6: Status 404 returned error can't find the container with id 11138c0ccf75f4e3f44c04a831a47ce0eab6ea00a1426987234c898575d081f6 Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.096867 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4980c5d8-c614-4943-aff6-cb07dac7593f" path="/var/lib/kubelet/pods/4980c5d8-c614-4943-aff6-cb07dac7593f/volumes" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.097885 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b63477b-8171-4222-815b-16c7b5ba0e13" path="/var/lib/kubelet/pods/9b63477b-8171-4222-815b-16c7b5ba0e13/volumes" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.128993 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm"] Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.704387 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" event={"ID":"3bceec37-623a-4649-b751-d2ce4db6d764","Type":"ContainerStarted","Data":"45eda0bd954b6ffcde1d0f9f47c5f078e616bd207526a770e67e0392770b9c22"} Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.704459 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" event={"ID":"3bceec37-623a-4649-b751-d2ce4db6d764","Type":"ContainerStarted","Data":"11138c0ccf75f4e3f44c04a831a47ce0eab6ea00a1426987234c898575d081f6"} Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.705476 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.706134 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" event={"ID":"ff29c930-8476-47c0-b258-e7756f0e28fb","Type":"ContainerStarted","Data":"2abb95b8ece2309edc8695fddf72c3cbb7505b61375971e1eb725b97ad7e6f82"} Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.706199 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" event={"ID":"ff29c930-8476-47c0-b258-e7756f0e28fb","Type":"ContainerStarted","Data":"9094ed301612c2132d2d04081fd6663b4271e00945c610ba5d58db70ab80c985"} Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.706808 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.726045 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.757624 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" podStartSLOduration=1.757596852 podStartE2EDuration="1.757596852s" podCreationTimestamp="2025-10-08 06:44:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:44:06.738062396 +0000 UTC m=+749.372502156" watchObservedRunningTime="2025-10-08 06:44:06.757596852 +0000 UTC m=+749.392036592" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.760920 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-69fdb48b8b-hkxgm" podStartSLOduration=1.7609138340000001 podStartE2EDuration="1.760913834s" podCreationTimestamp="2025-10-08 06:44:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:44:06.757217442 +0000 UTC m=+749.391657192" watchObservedRunningTime="2025-10-08 06:44:06.760913834 +0000 UTC m=+749.395353574" Oct 08 06:44:06 crc kubenswrapper[4810]: I1008 06:44:06.841858 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f7b67b7bd-jk4dr" Oct 08 06:44:11 crc kubenswrapper[4810]: I1008 06:44:11.120925 4810 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.432302 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.433064 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.433132 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.434053 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a91a7b94b569323b6b5755a125c6be2ca2114d957f75fbc9389bdbf60792809"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.434158 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://2a91a7b94b569323b6b5755a125c6be2ca2114d957f75fbc9389bdbf60792809" gracePeriod=600 Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.764780 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bqqqb"] Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.766995 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.785915 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bqqqb"] Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.814274 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="2a91a7b94b569323b6b5755a125c6be2ca2114d957f75fbc9389bdbf60792809" exitCode=0 Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.814369 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"2a91a7b94b569323b6b5755a125c6be2ca2114d957f75fbc9389bdbf60792809"} Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.814455 4810 scope.go:117] "RemoveContainer" containerID="f85489310335fb08e0e9aeb0f5009d545bd1d41a5c6acf3db9afeeed7e8e5528" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.913689 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-utilities\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.913763 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-catalog-content\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:19 crc kubenswrapper[4810]: I1008 06:44:19.913811 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mxqx\" (UniqueName: \"kubernetes.io/projected/9a885ad8-4441-41d5-a702-710a8d8cde63-kube-api-access-9mxqx\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.015126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-utilities\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.015212 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-catalog-content\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.015263 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mxqx\" (UniqueName: \"kubernetes.io/projected/9a885ad8-4441-41d5-a702-710a8d8cde63-kube-api-access-9mxqx\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.015795 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-utilities\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.016183 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-catalog-content\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.046619 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mxqx\" (UniqueName: \"kubernetes.io/projected/9a885ad8-4441-41d5-a702-710a8d8cde63-kube-api-access-9mxqx\") pod \"community-operators-bqqqb\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.086443 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.523665 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-85f4bbfddc-bx2db" Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.585484 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bqqqb"] Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.823995 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"bf6102000b45e1894934eca5fb9ef2a49eefef58f2ba5b5554021b8f1388a9d9"} Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.826119 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerID="d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133" exitCode=0 Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.826214 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqqqb" event={"ID":"9a885ad8-4441-41d5-a702-710a8d8cde63","Type":"ContainerDied","Data":"d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133"} Oct 08 06:44:20 crc kubenswrapper[4810]: I1008 06:44:20.826266 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqqqb" event={"ID":"9a885ad8-4441-41d5-a702-710a8d8cde63","Type":"ContainerStarted","Data":"3b88dd727ae8a3ac87bd7e9f115605fdfe8dc6c0b9d0b829c5215eca620f61f4"} Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.382909 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm"] Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.384344 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.387220 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ttq28" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.387228 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.392938 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ncsk2"] Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.396350 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.398150 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.400224 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.404555 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm"] Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.440814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cllg9\" (UniqueName: \"kubernetes.io/projected/41795b92-c731-421a-bf39-f3d48e99b84c-kube-api-access-cllg9\") pod \"frr-k8s-webhook-server-64bf5d555-qh2dm\" (UID: \"41795b92-c731-421a-bf39-f3d48e99b84c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.440914 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41795b92-c731-421a-bf39-f3d48e99b84c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qh2dm\" (UID: \"41795b92-c731-421a-bf39-f3d48e99b84c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.484866 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-ffrq4"] Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.486080 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.488520 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.488645 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.488799 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8ltzx" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.488524 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.509336 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-4xvt4"] Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.511087 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.516159 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.524109 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-4xvt4"] Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542045 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41795b92-c731-421a-bf39-f3d48e99b84c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qh2dm\" (UID: \"41795b92-c731-421a-bf39-f3d48e99b84c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542115 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-reloader\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542139 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c32cd68-ec2a-4134-a7f4-da9466733bc1-metrics-certs\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542160 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metrics-certs\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542181 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-sockets\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542202 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-metrics\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: E1008 06:44:21.542211 4810 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 08 06:44:21 crc kubenswrapper[4810]: E1008 06:44:21.542323 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41795b92-c731-421a-bf39-f3d48e99b84c-cert podName:41795b92-c731-421a-bf39-f3d48e99b84c nodeName:}" failed. No retries permitted until 2025-10-08 06:44:22.042302293 +0000 UTC m=+764.676742033 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41795b92-c731-421a-bf39-f3d48e99b84c-cert") pod "frr-k8s-webhook-server-64bf5d555-qh2dm" (UID: "41795b92-c731-421a-bf39-f3d48e99b84c") : secret "frr-k8s-webhook-server-cert" not found Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542433 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542508 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-startup\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542679 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metallb-excludel2\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542775 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cllg9\" (UniqueName: \"kubernetes.io/projected/41795b92-c731-421a-bf39-f3d48e99b84c-kube-api-access-cllg9\") pod \"frr-k8s-webhook-server-64bf5d555-qh2dm\" (UID: \"41795b92-c731-421a-bf39-f3d48e99b84c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542831 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-conf\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542857 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjvwg\" (UniqueName: \"kubernetes.io/projected/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-kube-api-access-bjvwg\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.542898 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6ng9\" (UniqueName: \"kubernetes.io/projected/6c32cd68-ec2a-4134-a7f4-da9466733bc1-kube-api-access-b6ng9\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.562790 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cllg9\" (UniqueName: \"kubernetes.io/projected/41795b92-c731-421a-bf39-f3d48e99b84c-kube-api-access-cllg9\") pod \"frr-k8s-webhook-server-64bf5d555-qh2dm\" (UID: \"41795b92-c731-421a-bf39-f3d48e99b84c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.644473 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-reloader\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.644560 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c32cd68-ec2a-4134-a7f4-da9466733bc1-metrics-certs\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.644600 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metrics-certs\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.644648 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-sockets\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.644698 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-metrics\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.644739 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.644854 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-startup\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: E1008 06:44:21.644899 4810 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 08 06:44:21 crc kubenswrapper[4810]: E1008 06:44:21.645029 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metrics-certs podName:30febcf0-5a02-4efd-b2c7-f1fadc2f0f51 nodeName:}" failed. No retries permitted until 2025-10-08 06:44:22.145001029 +0000 UTC m=+764.779440769 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metrics-certs") pod "speaker-ffrq4" (UID: "30febcf0-5a02-4efd-b2c7-f1fadc2f0f51") : secret "speaker-certs-secret" not found Oct 08 06:44:21 crc kubenswrapper[4810]: E1008 06:44:21.645205 4810 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 06:44:21 crc kubenswrapper[4810]: E1008 06:44:21.645315 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist podName:30febcf0-5a02-4efd-b2c7-f1fadc2f0f51 nodeName:}" failed. No retries permitted until 2025-10-08 06:44:22.145281826 +0000 UTC m=+764.779721586 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist") pod "speaker-ffrq4" (UID: "30febcf0-5a02-4efd-b2c7-f1fadc2f0f51") : secret "metallb-memberlist" not found Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645404 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metallb-excludel2\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645448 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-cert\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645491 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-conf\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645526 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-reloader\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645552 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-metrics\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645526 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjvwg\" (UniqueName: \"kubernetes.io/projected/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-kube-api-access-bjvwg\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645662 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6ng9\" (UniqueName: \"kubernetes.io/projected/6c32cd68-ec2a-4134-a7f4-da9466733bc1-kube-api-access-b6ng9\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645686 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-sockets\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645718 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn2q2\" (UniqueName: \"kubernetes.io/projected/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-kube-api-access-kn2q2\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645838 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-metrics-certs\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.645836 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-conf\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.646274 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metallb-excludel2\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.646400 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6c32cd68-ec2a-4134-a7f4-da9466733bc1-frr-startup\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.653440 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6c32cd68-ec2a-4134-a7f4-da9466733bc1-metrics-certs\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.663080 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6ng9\" (UniqueName: \"kubernetes.io/projected/6c32cd68-ec2a-4134-a7f4-da9466733bc1-kube-api-access-b6ng9\") pod \"frr-k8s-ncsk2\" (UID: \"6c32cd68-ec2a-4134-a7f4-da9466733bc1\") " pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.663204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjvwg\" (UniqueName: \"kubernetes.io/projected/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-kube-api-access-bjvwg\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.722283 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.747981 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-cert\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.748351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn2q2\" (UniqueName: \"kubernetes.io/projected/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-kube-api-access-kn2q2\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.748516 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-metrics-certs\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.754079 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-metrics-certs\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.760909 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-cert\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.771903 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn2q2\" (UniqueName: \"kubernetes.io/projected/f95f0a34-32e5-4bd3-a2f7-73afabbfc182-kube-api-access-kn2q2\") pod \"controller-68d546b9d8-4xvt4\" (UID: \"f95f0a34-32e5-4bd3-a2f7-73afabbfc182\") " pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.824825 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:21 crc kubenswrapper[4810]: I1008 06:44:21.844822 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqqqb" event={"ID":"9a885ad8-4441-41d5-a702-710a8d8cde63","Type":"ContainerStarted","Data":"11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b"} Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.053400 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41795b92-c731-421a-bf39-f3d48e99b84c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qh2dm\" (UID: \"41795b92-c731-421a-bf39-f3d48e99b84c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.058518 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41795b92-c731-421a-bf39-f3d48e99b84c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qh2dm\" (UID: \"41795b92-c731-421a-bf39-f3d48e99b84c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.158178 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metrics-certs\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.158319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:22 crc kubenswrapper[4810]: E1008 06:44:22.158586 4810 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 06:44:22 crc kubenswrapper[4810]: E1008 06:44:22.158750 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist podName:30febcf0-5a02-4efd-b2c7-f1fadc2f0f51 nodeName:}" failed. No retries permitted until 2025-10-08 06:44:23.15871291 +0000 UTC m=+765.793152670 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist") pod "speaker-ffrq4" (UID: "30febcf0-5a02-4efd-b2c7-f1fadc2f0f51") : secret "metallb-memberlist" not found Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.168227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-metrics-certs\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.275305 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-4xvt4"] Oct 08 06:44:22 crc kubenswrapper[4810]: W1008 06:44:22.285401 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf95f0a34_32e5_4bd3_a2f7_73afabbfc182.slice/crio-296d954d0c996f8f6e9feeff521465a69347c55b1073f12b09470615f62522db WatchSource:0}: Error finding container 296d954d0c996f8f6e9feeff521465a69347c55b1073f12b09470615f62522db: Status 404 returned error can't find the container with id 296d954d0c996f8f6e9feeff521465a69347c55b1073f12b09470615f62522db Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.311316 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.740185 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm"] Oct 08 06:44:22 crc kubenswrapper[4810]: W1008 06:44:22.748451 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41795b92_c731_421a_bf39_f3d48e99b84c.slice/crio-dd4d4ab4a68fcafc46686ce5a3244fdffa93cd65a344d6770775512bfb373944 WatchSource:0}: Error finding container dd4d4ab4a68fcafc46686ce5a3244fdffa93cd65a344d6770775512bfb373944: Status 404 returned error can't find the container with id dd4d4ab4a68fcafc46686ce5a3244fdffa93cd65a344d6770775512bfb373944 Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.854109 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4xvt4" event={"ID":"f95f0a34-32e5-4bd3-a2f7-73afabbfc182","Type":"ContainerStarted","Data":"8d42f7502721b09acb91e6df91262b1d00def17de6c15c13448b0ca8eb7fbdf4"} Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.854173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4xvt4" event={"ID":"f95f0a34-32e5-4bd3-a2f7-73afabbfc182","Type":"ContainerStarted","Data":"c2aecf263be265baae99df90bead1914a29f3b9d3f0c980c2b2053c83a1cff22"} Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.854189 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-4xvt4" event={"ID":"f95f0a34-32e5-4bd3-a2f7-73afabbfc182","Type":"ContainerStarted","Data":"296d954d0c996f8f6e9feeff521465a69347c55b1073f12b09470615f62522db"} Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.854267 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.856533 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerID="11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b" exitCode=0 Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.856612 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqqqb" event={"ID":"9a885ad8-4441-41d5-a702-710a8d8cde63","Type":"ContainerDied","Data":"11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b"} Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.858477 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" event={"ID":"41795b92-c731-421a-bf39-f3d48e99b84c","Type":"ContainerStarted","Data":"dd4d4ab4a68fcafc46686ce5a3244fdffa93cd65a344d6770775512bfb373944"} Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.859733 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerStarted","Data":"47c91de9566a6859c2ec08474d2e26bed003251f8d2003047ec1b63a3181796a"} Oct 08 06:44:22 crc kubenswrapper[4810]: I1008 06:44:22.882922 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-4xvt4" podStartSLOduration=1.882903362 podStartE2EDuration="1.882903362s" podCreationTimestamp="2025-10-08 06:44:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:44:22.879409895 +0000 UTC m=+765.513849635" watchObservedRunningTime="2025-10-08 06:44:22.882903362 +0000 UTC m=+765.517343102" Oct 08 06:44:23 crc kubenswrapper[4810]: I1008 06:44:23.175121 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:23 crc kubenswrapper[4810]: I1008 06:44:23.185554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/30febcf0-5a02-4efd-b2c7-f1fadc2f0f51-memberlist\") pod \"speaker-ffrq4\" (UID: \"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51\") " pod="metallb-system/speaker-ffrq4" Oct 08 06:44:23 crc kubenswrapper[4810]: I1008 06:44:23.304397 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ffrq4" Oct 08 06:44:23 crc kubenswrapper[4810]: W1008 06:44:23.404302 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30febcf0_5a02_4efd_b2c7_f1fadc2f0f51.slice/crio-289f6b6c4a32ef1e2e8e37e2257208c0f57463eb02da36f687dc56d1038e5c59 WatchSource:0}: Error finding container 289f6b6c4a32ef1e2e8e37e2257208c0f57463eb02da36f687dc56d1038e5c59: Status 404 returned error can't find the container with id 289f6b6c4a32ef1e2e8e37e2257208c0f57463eb02da36f687dc56d1038e5c59 Oct 08 06:44:23 crc kubenswrapper[4810]: I1008 06:44:23.871544 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqqqb" event={"ID":"9a885ad8-4441-41d5-a702-710a8d8cde63","Type":"ContainerStarted","Data":"68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1"} Oct 08 06:44:23 crc kubenswrapper[4810]: I1008 06:44:23.874376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ffrq4" event={"ID":"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51","Type":"ContainerStarted","Data":"82825122d3962af8596c4e2a0b1dd3bc97124161954b6b5be049b6eef20c751b"} Oct 08 06:44:23 crc kubenswrapper[4810]: I1008 06:44:23.874400 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ffrq4" event={"ID":"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51","Type":"ContainerStarted","Data":"289f6b6c4a32ef1e2e8e37e2257208c0f57463eb02da36f687dc56d1038e5c59"} Oct 08 06:44:24 crc kubenswrapper[4810]: I1008 06:44:24.886105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ffrq4" event={"ID":"30febcf0-5a02-4efd-b2c7-f1fadc2f0f51","Type":"ContainerStarted","Data":"6c5325a2e2398c381c1af62d3815ea6286398208bb5f4de47393b9848401d09c"} Oct 08 06:44:24 crc kubenswrapper[4810]: I1008 06:44:24.886857 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-ffrq4" Oct 08 06:44:24 crc kubenswrapper[4810]: I1008 06:44:24.907570 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bqqqb" podStartSLOduration=3.403473029 podStartE2EDuration="5.907547269s" podCreationTimestamp="2025-10-08 06:44:19 +0000 UTC" firstStartedPulling="2025-10-08 06:44:20.828652991 +0000 UTC m=+763.463092751" lastFinishedPulling="2025-10-08 06:44:23.332727241 +0000 UTC m=+765.967166991" observedRunningTime="2025-10-08 06:44:23.905324907 +0000 UTC m=+766.539764667" watchObservedRunningTime="2025-10-08 06:44:24.907547269 +0000 UTC m=+767.541987009" Oct 08 06:44:24 crc kubenswrapper[4810]: I1008 06:44:24.910062 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-ffrq4" podStartSLOduration=3.909954825 podStartE2EDuration="3.909954825s" podCreationTimestamp="2025-10-08 06:44:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:44:24.905862964 +0000 UTC m=+767.540302704" watchObservedRunningTime="2025-10-08 06:44:24.909954825 +0000 UTC m=+767.544394565" Oct 08 06:44:30 crc kubenswrapper[4810]: I1008 06:44:30.087396 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:30 crc kubenswrapper[4810]: I1008 06:44:30.088068 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:30 crc kubenswrapper[4810]: I1008 06:44:30.150026 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:31 crc kubenswrapper[4810]: I1008 06:44:31.055366 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:31 crc kubenswrapper[4810]: I1008 06:44:31.114816 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bqqqb"] Oct 08 06:44:31 crc kubenswrapper[4810]: I1008 06:44:31.968483 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" event={"ID":"41795b92-c731-421a-bf39-f3d48e99b84c","Type":"ContainerStarted","Data":"81a596fd488739070ea7d4508e5e6bf59f043b143ba026a6edd84e4b476f5d5a"} Oct 08 06:44:31 crc kubenswrapper[4810]: I1008 06:44:31.968996 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:31 crc kubenswrapper[4810]: I1008 06:44:31.971822 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c32cd68-ec2a-4134-a7f4-da9466733bc1" containerID="9fd605ed6bced22b654cb000d73c7b75666c31f3c3f21a49d35375bdb61b7610" exitCode=0 Oct 08 06:44:31 crc kubenswrapper[4810]: I1008 06:44:31.971934 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerDied","Data":"9fd605ed6bced22b654cb000d73c7b75666c31f3c3f21a49d35375bdb61b7610"} Oct 08 06:44:31 crc kubenswrapper[4810]: I1008 06:44:31.996887 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" podStartSLOduration=2.736033267 podStartE2EDuration="10.996844788s" podCreationTimestamp="2025-10-08 06:44:21 +0000 UTC" firstStartedPulling="2025-10-08 06:44:22.75261014 +0000 UTC m=+765.387049900" lastFinishedPulling="2025-10-08 06:44:31.013421681 +0000 UTC m=+773.647861421" observedRunningTime="2025-10-08 06:44:31.993025284 +0000 UTC m=+774.627465064" watchObservedRunningTime="2025-10-08 06:44:31.996844788 +0000 UTC m=+774.631284558" Oct 08 06:44:32 crc kubenswrapper[4810]: I1008 06:44:32.985522 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c32cd68-ec2a-4134-a7f4-da9466733bc1" containerID="d3a5fa69e44001b2056f7978e80cef57203428c0ac22a026a94426c96245743b" exitCode=0 Oct 08 06:44:32 crc kubenswrapper[4810]: I1008 06:44:32.985624 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerDied","Data":"d3a5fa69e44001b2056f7978e80cef57203428c0ac22a026a94426c96245743b"} Oct 08 06:44:32 crc kubenswrapper[4810]: I1008 06:44:32.987495 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bqqqb" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="registry-server" containerID="cri-o://68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1" gracePeriod=2 Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.310035 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-ffrq4" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.612566 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.784360 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-utilities\") pod \"9a885ad8-4441-41d5-a702-710a8d8cde63\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.784601 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mxqx\" (UniqueName: \"kubernetes.io/projected/9a885ad8-4441-41d5-a702-710a8d8cde63-kube-api-access-9mxqx\") pod \"9a885ad8-4441-41d5-a702-710a8d8cde63\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.784663 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-catalog-content\") pod \"9a885ad8-4441-41d5-a702-710a8d8cde63\" (UID: \"9a885ad8-4441-41d5-a702-710a8d8cde63\") " Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.786350 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-utilities" (OuterVolumeSpecName: "utilities") pod "9a885ad8-4441-41d5-a702-710a8d8cde63" (UID: "9a885ad8-4441-41d5-a702-710a8d8cde63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.793478 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a885ad8-4441-41d5-a702-710a8d8cde63-kube-api-access-9mxqx" (OuterVolumeSpecName: "kube-api-access-9mxqx") pod "9a885ad8-4441-41d5-a702-710a8d8cde63" (UID: "9a885ad8-4441-41d5-a702-710a8d8cde63"). InnerVolumeSpecName "kube-api-access-9mxqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.853090 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a885ad8-4441-41d5-a702-710a8d8cde63" (UID: "9a885ad8-4441-41d5-a702-710a8d8cde63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.887324 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.887371 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mxqx\" (UniqueName: \"kubernetes.io/projected/9a885ad8-4441-41d5-a702-710a8d8cde63-kube-api-access-9mxqx\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.887384 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a885ad8-4441-41d5-a702-710a8d8cde63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.996207 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerID="68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1" exitCode=0 Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.996346 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqqqb" event={"ID":"9a885ad8-4441-41d5-a702-710a8d8cde63","Type":"ContainerDied","Data":"68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1"} Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.996425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bqqqb" event={"ID":"9a885ad8-4441-41d5-a702-710a8d8cde63","Type":"ContainerDied","Data":"3b88dd727ae8a3ac87bd7e9f115605fdfe8dc6c0b9d0b829c5215eca620f61f4"} Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.996451 4810 scope.go:117] "RemoveContainer" containerID="68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.996689 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bqqqb" Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.999465 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c32cd68-ec2a-4134-a7f4-da9466733bc1" containerID="39665ddb172727f499ae4ec29e0f187c8cfb7d76a746054fdb21925680385e23" exitCode=0 Oct 08 06:44:33 crc kubenswrapper[4810]: I1008 06:44:33.999540 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerDied","Data":"39665ddb172727f499ae4ec29e0f187c8cfb7d76a746054fdb21925680385e23"} Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.018686 4810 scope.go:117] "RemoveContainer" containerID="11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.054312 4810 scope.go:117] "RemoveContainer" containerID="d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.060653 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bqqqb"] Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.082791 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bqqqb"] Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.095121 4810 scope.go:117] "RemoveContainer" containerID="68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1" Oct 08 06:44:34 crc kubenswrapper[4810]: E1008 06:44:34.096233 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1\": container with ID starting with 68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1 not found: ID does not exist" containerID="68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.096306 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1"} err="failed to get container status \"68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1\": rpc error: code = NotFound desc = could not find container \"68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1\": container with ID starting with 68d0a9695351fc1933f3da3c8e35107f44aba3e6764e7fb78b566ab1a9822bf1 not found: ID does not exist" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.096353 4810 scope.go:117] "RemoveContainer" containerID="11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b" Oct 08 06:44:34 crc kubenswrapper[4810]: E1008 06:44:34.096780 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b\": container with ID starting with 11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b not found: ID does not exist" containerID="11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.096809 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b"} err="failed to get container status \"11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b\": rpc error: code = NotFound desc = could not find container \"11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b\": container with ID starting with 11eae7ed573a32ee47425a2e8ce8f3ee482ab64494f73e6968b125f7dbbb254b not found: ID does not exist" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.096831 4810 scope.go:117] "RemoveContainer" containerID="d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133" Oct 08 06:44:34 crc kubenswrapper[4810]: E1008 06:44:34.097400 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133\": container with ID starting with d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133 not found: ID does not exist" containerID="d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.097427 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133"} err="failed to get container status \"d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133\": rpc error: code = NotFound desc = could not find container \"d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133\": container with ID starting with d2856da31f76609780bb654c11a9cccae64e65f9205fd421fa424ed70bad9133 not found: ID does not exist" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.917404 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z"] Oct 08 06:44:34 crc kubenswrapper[4810]: E1008 06:44:34.918187 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="extract-utilities" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.918211 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="extract-utilities" Oct 08 06:44:34 crc kubenswrapper[4810]: E1008 06:44:34.918235 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="registry-server" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.918246 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="registry-server" Oct 08 06:44:34 crc kubenswrapper[4810]: E1008 06:44:34.918278 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="extract-content" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.918290 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="extract-content" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.918445 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" containerName="registry-server" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.920387 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.923779 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z"] Oct 08 06:44:34 crc kubenswrapper[4810]: I1008 06:44:34.958478 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.004727 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clk4s\" (UniqueName: \"kubernetes.io/projected/497ef608-88af-4370-8886-c5377ec5b1e8-kube-api-access-clk4s\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.004808 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.004844 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.016932 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerStarted","Data":"d59f41c6090f11b8d9b8064ca500dee0dc5d5f481713f625a106dfc486b09c90"} Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.016991 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerStarted","Data":"5bd912ebf77f21cfb5e79f83cd3a8a1ebc1f7b2e3306ea81cd4d78ef48002494"} Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.017002 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerStarted","Data":"c7c4745a7c6f050d68521f60ae4a2d3c73c060f970a325a613a4f48442419e1d"} Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.017011 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerStarted","Data":"56d14a43ccf81a8a036d5cb86339dd807e24c85515cf631d4173307d753e0493"} Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.017021 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerStarted","Data":"2064ec1de37166b8babf9291978b1fc91596ba1f886ba3aaac1ddaaba1fb5a64"} Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.106328 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clk4s\" (UniqueName: \"kubernetes.io/projected/497ef608-88af-4370-8886-c5377ec5b1e8-kube-api-access-clk4s\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.106411 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.106443 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.107536 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.107927 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.133520 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clk4s\" (UniqueName: \"kubernetes.io/projected/497ef608-88af-4370-8886-c5377ec5b1e8-kube-api-access-clk4s\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.277002 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:35 crc kubenswrapper[4810]: I1008 06:44:35.785813 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z"] Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.026694 4810 generic.go:334] "Generic (PLEG): container finished" podID="497ef608-88af-4370-8886-c5377ec5b1e8" containerID="1d96d529b17201fc15af74d967ec299778f969d56606d1938cd4e166a0ef6760" exitCode=0 Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.026833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" event={"ID":"497ef608-88af-4370-8886-c5377ec5b1e8","Type":"ContainerDied","Data":"1d96d529b17201fc15af74d967ec299778f969d56606d1938cd4e166a0ef6760"} Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.027391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" event={"ID":"497ef608-88af-4370-8886-c5377ec5b1e8","Type":"ContainerStarted","Data":"c2fa0ed8aff2a44355bdfc8ea80e902e8c15c12a32fa6132f9fe0267a8866888"} Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.036775 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ncsk2" event={"ID":"6c32cd68-ec2a-4134-a7f4-da9466733bc1","Type":"ContainerStarted","Data":"d10a0f6ebf71828057fe9e387b2361a9e9e47c381da456b816d65e2d7b80d242"} Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.037070 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.089129 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ncsk2" podStartSLOduration=6.03438313 podStartE2EDuration="15.089094253s" podCreationTimestamp="2025-10-08 06:44:21 +0000 UTC" firstStartedPulling="2025-10-08 06:44:21.920824139 +0000 UTC m=+764.555263879" lastFinishedPulling="2025-10-08 06:44:30.975535262 +0000 UTC m=+773.609975002" observedRunningTime="2025-10-08 06:44:36.07767725 +0000 UTC m=+778.712117010" watchObservedRunningTime="2025-10-08 06:44:36.089094253 +0000 UTC m=+778.723534013" Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.093366 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a885ad8-4441-41d5-a702-710a8d8cde63" path="/var/lib/kubelet/pods/9a885ad8-4441-41d5-a702-710a8d8cde63/volumes" Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.722582 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:36 crc kubenswrapper[4810]: I1008 06:44:36.758749 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.464117 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tbzq9"] Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.470940 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.473139 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tbzq9"] Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.562624 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.562708 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d29f\" (UniqueName: \"kubernetes.io/projected/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-kube-api-access-7d29f\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.562738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-utilities\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.663797 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d29f\" (UniqueName: \"kubernetes.io/projected/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-kube-api-access-7d29f\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.663856 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-utilities\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.663898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.664387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.665078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-utilities\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.683617 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d29f\" (UniqueName: \"kubernetes.io/projected/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-kube-api-access-7d29f\") pod \"redhat-operators-tbzq9\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.825494 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.855350 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-88f7b"] Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.856346 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.895274 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-88f7b"] Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.969469 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6psf\" (UniqueName: \"kubernetes.io/projected/f7215803-d955-4883-b990-a6fdfe514f70-kube-api-access-m6psf\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.969550 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-catalog-content\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:38 crc kubenswrapper[4810]: I1008 06:44:38.969694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-utilities\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.071052 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-utilities\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.071106 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6psf\" (UniqueName: \"kubernetes.io/projected/f7215803-d955-4883-b990-a6fdfe514f70-kube-api-access-m6psf\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.071128 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-catalog-content\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.071750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-catalog-content\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.071750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-utilities\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.089832 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6psf\" (UniqueName: \"kubernetes.io/projected/f7215803-d955-4883-b990-a6fdfe514f70-kube-api-access-m6psf\") pod \"redhat-marketplace-88f7b\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.176732 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:39 crc kubenswrapper[4810]: I1008 06:44:39.969281 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tbzq9"] Oct 08 06:44:39 crc kubenswrapper[4810]: W1008 06:44:39.974915 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode25e5fd0_12f0_4b49_9cc6_aac032a3b270.slice/crio-77e18b42eae5edc87674b1d22720e549901628ab6b89dca080234974714b2280 WatchSource:0}: Error finding container 77e18b42eae5edc87674b1d22720e549901628ab6b89dca080234974714b2280: Status 404 returned error can't find the container with id 77e18b42eae5edc87674b1d22720e549901628ab6b89dca080234974714b2280 Oct 08 06:44:40 crc kubenswrapper[4810]: I1008 06:44:40.073587 4810 generic.go:334] "Generic (PLEG): container finished" podID="497ef608-88af-4370-8886-c5377ec5b1e8" containerID="7fae48c63695233e5f1de48b6d344c4ae3e69428c15a42eafe305223e34cb634" exitCode=0 Oct 08 06:44:40 crc kubenswrapper[4810]: I1008 06:44:40.086389 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" event={"ID":"497ef608-88af-4370-8886-c5377ec5b1e8","Type":"ContainerDied","Data":"7fae48c63695233e5f1de48b6d344c4ae3e69428c15a42eafe305223e34cb634"} Oct 08 06:44:40 crc kubenswrapper[4810]: I1008 06:44:40.086425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbzq9" event={"ID":"e25e5fd0-12f0-4b49-9cc6-aac032a3b270","Type":"ContainerStarted","Data":"77e18b42eae5edc87674b1d22720e549901628ab6b89dca080234974714b2280"} Oct 08 06:44:40 crc kubenswrapper[4810]: I1008 06:44:40.243502 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-88f7b"] Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.089853 4810 generic.go:334] "Generic (PLEG): container finished" podID="497ef608-88af-4370-8886-c5377ec5b1e8" containerID="c0ac9fbfbb01b245952af95258f90b314264ce7e7e8e4643dc473a05434507b9" exitCode=0 Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.089997 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" event={"ID":"497ef608-88af-4370-8886-c5377ec5b1e8","Type":"ContainerDied","Data":"c0ac9fbfbb01b245952af95258f90b314264ce7e7e8e4643dc473a05434507b9"} Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.091906 4810 generic.go:334] "Generic (PLEG): container finished" podID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerID="066528a016f7b4f6d06b402231b6b44169060e34fea503879269c1f443245f56" exitCode=0 Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.092053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbzq9" event={"ID":"e25e5fd0-12f0-4b49-9cc6-aac032a3b270","Type":"ContainerDied","Data":"066528a016f7b4f6d06b402231b6b44169060e34fea503879269c1f443245f56"} Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.094209 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7215803-d955-4883-b990-a6fdfe514f70" containerID="d9ef68748b480421fb5724673e3379f843eb50a23b80b17454fefc14b9caa4d4" exitCode=0 Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.094256 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-88f7b" event={"ID":"f7215803-d955-4883-b990-a6fdfe514f70","Type":"ContainerDied","Data":"d9ef68748b480421fb5724673e3379f843eb50a23b80b17454fefc14b9caa4d4"} Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.094289 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-88f7b" event={"ID":"f7215803-d955-4883-b990-a6fdfe514f70","Type":"ContainerStarted","Data":"2e0978d9bda4d6b01ca0900dfb9bbe70b69c343f81ed78e896e9b69986b27cc6"} Oct 08 06:44:41 crc kubenswrapper[4810]: I1008 06:44:41.829140 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-4xvt4" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.102169 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbzq9" event={"ID":"e25e5fd0-12f0-4b49-9cc6-aac032a3b270","Type":"ContainerStarted","Data":"223dfe363327654ff5ec4a8b8cd0e476701c6406571a5c6393c2f0324ff9c1ce"} Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.104225 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7215803-d955-4883-b990-a6fdfe514f70" containerID="d448a7d74e4c8c07bd40e67280ad40e06793b39f0e227af2e04cf6e83068c355" exitCode=0 Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.104709 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-88f7b" event={"ID":"f7215803-d955-4883-b990-a6fdfe514f70","Type":"ContainerDied","Data":"d448a7d74e4c8c07bd40e67280ad40e06793b39f0e227af2e04cf6e83068c355"} Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.320714 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qh2dm" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.489282 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.648471 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-util\") pod \"497ef608-88af-4370-8886-c5377ec5b1e8\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.648821 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clk4s\" (UniqueName: \"kubernetes.io/projected/497ef608-88af-4370-8886-c5377ec5b1e8-kube-api-access-clk4s\") pod \"497ef608-88af-4370-8886-c5377ec5b1e8\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.648871 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-bundle\") pod \"497ef608-88af-4370-8886-c5377ec5b1e8\" (UID: \"497ef608-88af-4370-8886-c5377ec5b1e8\") " Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.649895 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-bundle" (OuterVolumeSpecName: "bundle") pod "497ef608-88af-4370-8886-c5377ec5b1e8" (UID: "497ef608-88af-4370-8886-c5377ec5b1e8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.654415 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/497ef608-88af-4370-8886-c5377ec5b1e8-kube-api-access-clk4s" (OuterVolumeSpecName: "kube-api-access-clk4s") pod "497ef608-88af-4370-8886-c5377ec5b1e8" (UID: "497ef608-88af-4370-8886-c5377ec5b1e8"). InnerVolumeSpecName "kube-api-access-clk4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.662628 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-util" (OuterVolumeSpecName: "util") pod "497ef608-88af-4370-8886-c5377ec5b1e8" (UID: "497ef608-88af-4370-8886-c5377ec5b1e8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.750695 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-util\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.750755 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clk4s\" (UniqueName: \"kubernetes.io/projected/497ef608-88af-4370-8886-c5377ec5b1e8-kube-api-access-clk4s\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:42 crc kubenswrapper[4810]: I1008 06:44:42.750771 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/497ef608-88af-4370-8886-c5377ec5b1e8-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:43 crc kubenswrapper[4810]: I1008 06:44:43.112948 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" event={"ID":"497ef608-88af-4370-8886-c5377ec5b1e8","Type":"ContainerDied","Data":"c2fa0ed8aff2a44355bdfc8ea80e902e8c15c12a32fa6132f9fe0267a8866888"} Oct 08 06:44:43 crc kubenswrapper[4810]: I1008 06:44:43.113036 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2fa0ed8aff2a44355bdfc8ea80e902e8c15c12a32fa6132f9fe0267a8866888" Oct 08 06:44:43 crc kubenswrapper[4810]: I1008 06:44:43.112997 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z" Oct 08 06:44:43 crc kubenswrapper[4810]: I1008 06:44:43.114727 4810 generic.go:334] "Generic (PLEG): container finished" podID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerID="223dfe363327654ff5ec4a8b8cd0e476701c6406571a5c6393c2f0324ff9c1ce" exitCode=0 Oct 08 06:44:43 crc kubenswrapper[4810]: I1008 06:44:43.114811 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbzq9" event={"ID":"e25e5fd0-12f0-4b49-9cc6-aac032a3b270","Type":"ContainerDied","Data":"223dfe363327654ff5ec4a8b8cd0e476701c6406571a5c6393c2f0324ff9c1ce"} Oct 08 06:44:43 crc kubenswrapper[4810]: I1008 06:44:43.119367 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-88f7b" event={"ID":"f7215803-d955-4883-b990-a6fdfe514f70","Type":"ContainerStarted","Data":"d8432053574ce7176fcb97dd048d231acb948584d69f8506f8ae1b245bb6a286"} Oct 08 06:44:43 crc kubenswrapper[4810]: I1008 06:44:43.175710 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-88f7b" podStartSLOduration=3.768686349 podStartE2EDuration="5.175685736s" podCreationTimestamp="2025-10-08 06:44:38 +0000 UTC" firstStartedPulling="2025-10-08 06:44:41.096161754 +0000 UTC m=+783.730601534" lastFinishedPulling="2025-10-08 06:44:42.503161181 +0000 UTC m=+785.137600921" observedRunningTime="2025-10-08 06:44:43.172863769 +0000 UTC m=+785.807303549" watchObservedRunningTime="2025-10-08 06:44:43.175685736 +0000 UTC m=+785.810125486" Oct 08 06:44:44 crc kubenswrapper[4810]: I1008 06:44:44.127670 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbzq9" event={"ID":"e25e5fd0-12f0-4b49-9cc6-aac032a3b270","Type":"ContainerStarted","Data":"bb1c6539055ec91f99bdf696636082bebc5ed1e6ee7f07a9cb3ca422559d80fd"} Oct 08 06:44:44 crc kubenswrapper[4810]: I1008 06:44:44.154168 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tbzq9" podStartSLOduration=3.629913354 podStartE2EDuration="6.154144028s" podCreationTimestamp="2025-10-08 06:44:38 +0000 UTC" firstStartedPulling="2025-10-08 06:44:41.094516868 +0000 UTC m=+783.728956648" lastFinishedPulling="2025-10-08 06:44:43.618747582 +0000 UTC m=+786.253187322" observedRunningTime="2025-10-08 06:44:44.147693591 +0000 UTC m=+786.782133351" watchObservedRunningTime="2025-10-08 06:44:44.154144028 +0000 UTC m=+786.788583768" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.539483 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49"] Oct 08 06:44:46 crc kubenswrapper[4810]: E1008 06:44:46.539744 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="497ef608-88af-4370-8886-c5377ec5b1e8" containerName="pull" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.539758 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="497ef608-88af-4370-8886-c5377ec5b1e8" containerName="pull" Oct 08 06:44:46 crc kubenswrapper[4810]: E1008 06:44:46.539771 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="497ef608-88af-4370-8886-c5377ec5b1e8" containerName="util" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.539778 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="497ef608-88af-4370-8886-c5377ec5b1e8" containerName="util" Oct 08 06:44:46 crc kubenswrapper[4810]: E1008 06:44:46.539792 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="497ef608-88af-4370-8886-c5377ec5b1e8" containerName="extract" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.539798 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="497ef608-88af-4370-8886-c5377ec5b1e8" containerName="extract" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.539908 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="497ef608-88af-4370-8886-c5377ec5b1e8" containerName="extract" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.540412 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.542855 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.542919 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-4x7zx" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.544748 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.576898 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49"] Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.704275 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-457q5\" (UniqueName: \"kubernetes.io/projected/8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9-kube-api-access-457q5\") pod \"cert-manager-operator-controller-manager-57cd46d6d-p9t49\" (UID: \"8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.806232 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-457q5\" (UniqueName: \"kubernetes.io/projected/8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9-kube-api-access-457q5\") pod \"cert-manager-operator-controller-manager-57cd46d6d-p9t49\" (UID: \"8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.852720 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-457q5\" (UniqueName: \"kubernetes.io/projected/8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9-kube-api-access-457q5\") pod \"cert-manager-operator-controller-manager-57cd46d6d-p9t49\" (UID: \"8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" Oct 08 06:44:46 crc kubenswrapper[4810]: I1008 06:44:46.865290 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" Oct 08 06:44:47 crc kubenswrapper[4810]: I1008 06:44:47.358799 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49"] Oct 08 06:44:47 crc kubenswrapper[4810]: W1008 06:44:47.373891 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8eaaf600_32e2_4cca_bdd2_ee1a70af0dd9.slice/crio-2cfb148995e54682d48584fd1d1af34d534c017b8df2acfe5f7eba0250ed5ceb WatchSource:0}: Error finding container 2cfb148995e54682d48584fd1d1af34d534c017b8df2acfe5f7eba0250ed5ceb: Status 404 returned error can't find the container with id 2cfb148995e54682d48584fd1d1af34d534c017b8df2acfe5f7eba0250ed5ceb Oct 08 06:44:48 crc kubenswrapper[4810]: I1008 06:44:48.153044 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" event={"ID":"8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9","Type":"ContainerStarted","Data":"2cfb148995e54682d48584fd1d1af34d534c017b8df2acfe5f7eba0250ed5ceb"} Oct 08 06:44:48 crc kubenswrapper[4810]: I1008 06:44:48.826315 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:48 crc kubenswrapper[4810]: I1008 06:44:48.826936 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:49 crc kubenswrapper[4810]: I1008 06:44:49.177370 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:49 crc kubenswrapper[4810]: I1008 06:44:49.177617 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:49 crc kubenswrapper[4810]: I1008 06:44:49.228123 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:49 crc kubenswrapper[4810]: I1008 06:44:49.878819 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tbzq9" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="registry-server" probeResult="failure" output=< Oct 08 06:44:49 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Oct 08 06:44:49 crc kubenswrapper[4810]: > Oct 08 06:44:50 crc kubenswrapper[4810]: I1008 06:44:50.222648 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:51 crc kubenswrapper[4810]: I1008 06:44:51.726206 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ncsk2" Oct 08 06:44:51 crc kubenswrapper[4810]: I1008 06:44:51.853845 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-88f7b"] Oct 08 06:44:53 crc kubenswrapper[4810]: I1008 06:44:53.192656 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-88f7b" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="registry-server" containerID="cri-o://d8432053574ce7176fcb97dd048d231acb948584d69f8506f8ae1b245bb6a286" gracePeriod=2 Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.202658 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7215803-d955-4883-b990-a6fdfe514f70" containerID="d8432053574ce7176fcb97dd048d231acb948584d69f8506f8ae1b245bb6a286" exitCode=0 Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.202759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-88f7b" event={"ID":"f7215803-d955-4883-b990-a6fdfe514f70","Type":"ContainerDied","Data":"d8432053574ce7176fcb97dd048d231acb948584d69f8506f8ae1b245bb6a286"} Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.752984 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.845780 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-utilities\") pod \"f7215803-d955-4883-b990-a6fdfe514f70\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.845946 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-catalog-content\") pod \"f7215803-d955-4883-b990-a6fdfe514f70\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.846015 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6psf\" (UniqueName: \"kubernetes.io/projected/f7215803-d955-4883-b990-a6fdfe514f70-kube-api-access-m6psf\") pod \"f7215803-d955-4883-b990-a6fdfe514f70\" (UID: \"f7215803-d955-4883-b990-a6fdfe514f70\") " Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.848684 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-utilities" (OuterVolumeSpecName: "utilities") pod "f7215803-d955-4883-b990-a6fdfe514f70" (UID: "f7215803-d955-4883-b990-a6fdfe514f70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.856846 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7215803-d955-4883-b990-a6fdfe514f70-kube-api-access-m6psf" (OuterVolumeSpecName: "kube-api-access-m6psf") pod "f7215803-d955-4883-b990-a6fdfe514f70" (UID: "f7215803-d955-4883-b990-a6fdfe514f70"). InnerVolumeSpecName "kube-api-access-m6psf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.864589 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7215803-d955-4883-b990-a6fdfe514f70" (UID: "f7215803-d955-4883-b990-a6fdfe514f70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.947145 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.947183 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6psf\" (UniqueName: \"kubernetes.io/projected/f7215803-d955-4883-b990-a6fdfe514f70-kube-api-access-m6psf\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:54 crc kubenswrapper[4810]: I1008 06:44:54.947196 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7215803-d955-4883-b990-a6fdfe514f70-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.209946 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" event={"ID":"8eaaf600-32e2-4cca-bdd2-ee1a70af0dd9","Type":"ContainerStarted","Data":"8fc71b3feb8c58834070acaab3f1337299a36ba7ba5ccee3fa2f905345bf233f"} Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.211706 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-88f7b" event={"ID":"f7215803-d955-4883-b990-a6fdfe514f70","Type":"ContainerDied","Data":"2e0978d9bda4d6b01ca0900dfb9bbe70b69c343f81ed78e896e9b69986b27cc6"} Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.211745 4810 scope.go:117] "RemoveContainer" containerID="d8432053574ce7176fcb97dd048d231acb948584d69f8506f8ae1b245bb6a286" Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.211752 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-88f7b" Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.225360 4810 scope.go:117] "RemoveContainer" containerID="d448a7d74e4c8c07bd40e67280ad40e06793b39f0e227af2e04cf6e83068c355" Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.262392 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-p9t49" podStartSLOduration=2.104844572 podStartE2EDuration="9.262364191s" podCreationTimestamp="2025-10-08 06:44:46 +0000 UTC" firstStartedPulling="2025-10-08 06:44:47.376202059 +0000 UTC m=+790.010641789" lastFinishedPulling="2025-10-08 06:44:54.533721668 +0000 UTC m=+797.168161408" observedRunningTime="2025-10-08 06:44:55.232648786 +0000 UTC m=+797.867088526" watchObservedRunningTime="2025-10-08 06:44:55.262364191 +0000 UTC m=+797.896803931" Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.267102 4810 scope.go:117] "RemoveContainer" containerID="d9ef68748b480421fb5724673e3379f843eb50a23b80b17454fefc14b9caa4d4" Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.274439 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-88f7b"] Oct 08 06:44:55 crc kubenswrapper[4810]: I1008 06:44:55.278137 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-88f7b"] Oct 08 06:44:56 crc kubenswrapper[4810]: I1008 06:44:56.080506 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7215803-d955-4883-b990-a6fdfe514f70" path="/var/lib/kubelet/pods/f7215803-d955-4883-b990-a6fdfe514f70/volumes" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.523876 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-7chth"] Oct 08 06:44:57 crc kubenswrapper[4810]: E1008 06:44:57.524589 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="registry-server" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.524608 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="registry-server" Oct 08 06:44:57 crc kubenswrapper[4810]: E1008 06:44:57.524634 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="extract-utilities" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.524646 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="extract-utilities" Oct 08 06:44:57 crc kubenswrapper[4810]: E1008 06:44:57.524675 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="extract-content" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.524687 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="extract-content" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.524866 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7215803-d955-4883-b990-a6fdfe514f70" containerName="registry-server" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.525495 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.527168 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-7qg4d" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.528613 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.530467 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.537478 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-7chth"] Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.683018 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwxsb\" (UniqueName: \"kubernetes.io/projected/0d166b56-ad1e-44dc-96b3-58d1d3116642-kube-api-access-gwxsb\") pod \"cert-manager-webhook-d969966f-7chth\" (UID: \"0d166b56-ad1e-44dc-96b3-58d1d3116642\") " pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.683279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d166b56-ad1e-44dc-96b3-58d1d3116642-bound-sa-token\") pod \"cert-manager-webhook-d969966f-7chth\" (UID: \"0d166b56-ad1e-44dc-96b3-58d1d3116642\") " pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.784615 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwxsb\" (UniqueName: \"kubernetes.io/projected/0d166b56-ad1e-44dc-96b3-58d1d3116642-kube-api-access-gwxsb\") pod \"cert-manager-webhook-d969966f-7chth\" (UID: \"0d166b56-ad1e-44dc-96b3-58d1d3116642\") " pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.785024 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d166b56-ad1e-44dc-96b3-58d1d3116642-bound-sa-token\") pod \"cert-manager-webhook-d969966f-7chth\" (UID: \"0d166b56-ad1e-44dc-96b3-58d1d3116642\") " pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.813564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwxsb\" (UniqueName: \"kubernetes.io/projected/0d166b56-ad1e-44dc-96b3-58d1d3116642-kube-api-access-gwxsb\") pod \"cert-manager-webhook-d969966f-7chth\" (UID: \"0d166b56-ad1e-44dc-96b3-58d1d3116642\") " pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.814368 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0d166b56-ad1e-44dc-96b3-58d1d3116642-bound-sa-token\") pod \"cert-manager-webhook-d969966f-7chth\" (UID: \"0d166b56-ad1e-44dc-96b3-58d1d3116642\") " pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:57 crc kubenswrapper[4810]: I1008 06:44:57.849888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:44:58 crc kubenswrapper[4810]: I1008 06:44:58.306204 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-7chth"] Oct 08 06:44:58 crc kubenswrapper[4810]: I1008 06:44:58.873655 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:58 crc kubenswrapper[4810]: I1008 06:44:58.917046 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:44:59 crc kubenswrapper[4810]: I1008 06:44:59.241172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-7chth" event={"ID":"0d166b56-ad1e-44dc-96b3-58d1d3116642","Type":"ContainerStarted","Data":"56e1d61eed233670e75e221d9fc560f192f2207a18cccf33429f9b023f17f172"} Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.130914 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67"] Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.131700 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.134348 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.134700 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.144657 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67"] Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.221456 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c521b14-3481-4326-9219-ec19d2b6099b-secret-volume\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.221802 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84vh4\" (UniqueName: \"kubernetes.io/projected/7c521b14-3481-4326-9219-ec19d2b6099b-kube-api-access-84vh4\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.221824 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c521b14-3481-4326-9219-ec19d2b6099b-config-volume\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.322872 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c521b14-3481-4326-9219-ec19d2b6099b-secret-volume\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.322991 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c521b14-3481-4326-9219-ec19d2b6099b-config-volume\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.323021 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84vh4\" (UniqueName: \"kubernetes.io/projected/7c521b14-3481-4326-9219-ec19d2b6099b-kube-api-access-84vh4\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.325017 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c521b14-3481-4326-9219-ec19d2b6099b-config-volume\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.329151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c521b14-3481-4326-9219-ec19d2b6099b-secret-volume\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.343748 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84vh4\" (UniqueName: \"kubernetes.io/projected/7c521b14-3481-4326-9219-ec19d2b6099b-kube-api-access-84vh4\") pod \"collect-profiles-29331765-ttj67\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.500505 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:00 crc kubenswrapper[4810]: I1008 06:45:00.929343 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67"] Oct 08 06:45:01 crc kubenswrapper[4810]: I1008 06:45:01.263273 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" event={"ID":"7c521b14-3481-4326-9219-ec19d2b6099b","Type":"ContainerStarted","Data":"5daba72877a3db4832410476f48d6f15beee2f2205e16ff39b64326ce3524c11"} Oct 08 06:45:01 crc kubenswrapper[4810]: I1008 06:45:01.263348 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" event={"ID":"7c521b14-3481-4326-9219-ec19d2b6099b","Type":"ContainerStarted","Data":"c7eac283dd6170ce048f1f698e5bd389a5c964978fe8ad47305cedc3cef00e6c"} Oct 08 06:45:01 crc kubenswrapper[4810]: I1008 06:45:01.291480 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" podStartSLOduration=1.291460717 podStartE2EDuration="1.291460717s" podCreationTimestamp="2025-10-08 06:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:45:01.289420051 +0000 UTC m=+803.923859791" watchObservedRunningTime="2025-10-08 06:45:01.291460717 +0000 UTC m=+803.925900458" Oct 08 06:45:01 crc kubenswrapper[4810]: I1008 06:45:01.449228 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tbzq9"] Oct 08 06:45:01 crc kubenswrapper[4810]: I1008 06:45:01.450081 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tbzq9" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="registry-server" containerID="cri-o://bb1c6539055ec91f99bdf696636082bebc5ed1e6ee7f07a9cb3ca422559d80fd" gracePeriod=2 Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.143094 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r"] Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.143839 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.146149 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-92dcp" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.163568 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r"] Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.253610 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/82648b77-ff5f-4c3b-93b9-3a860b9f4149-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-cfv7r\" (UID: \"82648b77-ff5f-4c3b-93b9-3a860b9f4149\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.253694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p64j2\" (UniqueName: \"kubernetes.io/projected/82648b77-ff5f-4c3b-93b9-3a860b9f4149-kube-api-access-p64j2\") pod \"cert-manager-cainjector-7d9f95dbf-cfv7r\" (UID: \"82648b77-ff5f-4c3b-93b9-3a860b9f4149\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.299952 4810 generic.go:334] "Generic (PLEG): container finished" podID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerID="bb1c6539055ec91f99bdf696636082bebc5ed1e6ee7f07a9cb3ca422559d80fd" exitCode=0 Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.300050 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbzq9" event={"ID":"e25e5fd0-12f0-4b49-9cc6-aac032a3b270","Type":"ContainerDied","Data":"bb1c6539055ec91f99bdf696636082bebc5ed1e6ee7f07a9cb3ca422559d80fd"} Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.303320 4810 generic.go:334] "Generic (PLEG): container finished" podID="7c521b14-3481-4326-9219-ec19d2b6099b" containerID="5daba72877a3db4832410476f48d6f15beee2f2205e16ff39b64326ce3524c11" exitCode=0 Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.303395 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" event={"ID":"7c521b14-3481-4326-9219-ec19d2b6099b","Type":"ContainerDied","Data":"5daba72877a3db4832410476f48d6f15beee2f2205e16ff39b64326ce3524c11"} Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.356215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/82648b77-ff5f-4c3b-93b9-3a860b9f4149-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-cfv7r\" (UID: \"82648b77-ff5f-4c3b-93b9-3a860b9f4149\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.356291 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p64j2\" (UniqueName: \"kubernetes.io/projected/82648b77-ff5f-4c3b-93b9-3a860b9f4149-kube-api-access-p64j2\") pod \"cert-manager-cainjector-7d9f95dbf-cfv7r\" (UID: \"82648b77-ff5f-4c3b-93b9-3a860b9f4149\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.377730 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p64j2\" (UniqueName: \"kubernetes.io/projected/82648b77-ff5f-4c3b-93b9-3a860b9f4149-kube-api-access-p64j2\") pod \"cert-manager-cainjector-7d9f95dbf-cfv7r\" (UID: \"82648b77-ff5f-4c3b-93b9-3a860b9f4149\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.399877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/82648b77-ff5f-4c3b-93b9-3a860b9f4149-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-cfv7r\" (UID: \"82648b77-ff5f-4c3b-93b9-3a860b9f4149\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:02 crc kubenswrapper[4810]: I1008 06:45:02.468887 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.700884 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r"] Oct 08 06:45:03 crc kubenswrapper[4810]: W1008 06:45:03.711820 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82648b77_ff5f_4c3b_93b9_3a860b9f4149.slice/crio-0cf64bf3f947d1eec413dbc6d59f72718678687aa211a677e11b3dba8cc1d5db WatchSource:0}: Error finding container 0cf64bf3f947d1eec413dbc6d59f72718678687aa211a677e11b3dba8cc1d5db: Status 404 returned error can't find the container with id 0cf64bf3f947d1eec413dbc6d59f72718678687aa211a677e11b3dba8cc1d5db Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.757160 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.840271 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.880615 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d29f\" (UniqueName: \"kubernetes.io/projected/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-kube-api-access-7d29f\") pod \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.880667 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content\") pod \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.880892 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-utilities\") pod \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.882866 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-utilities" (OuterVolumeSpecName: "utilities") pod "e25e5fd0-12f0-4b49-9cc6-aac032a3b270" (UID: "e25e5fd0-12f0-4b49-9cc6-aac032a3b270"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.903275 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-kube-api-access-7d29f" (OuterVolumeSpecName: "kube-api-access-7d29f") pod "e25e5fd0-12f0-4b49-9cc6-aac032a3b270" (UID: "e25e5fd0-12f0-4b49-9cc6-aac032a3b270"). InnerVolumeSpecName "kube-api-access-7d29f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.989797 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e25e5fd0-12f0-4b49-9cc6-aac032a3b270" (UID: "e25e5fd0-12f0-4b49-9cc6-aac032a3b270"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.995702 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c521b14-3481-4326-9219-ec19d2b6099b-config-volume\") pod \"7c521b14-3481-4326-9219-ec19d2b6099b\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.995885 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c521b14-3481-4326-9219-ec19d2b6099b-secret-volume\") pod \"7c521b14-3481-4326-9219-ec19d2b6099b\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.996093 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84vh4\" (UniqueName: \"kubernetes.io/projected/7c521b14-3481-4326-9219-ec19d2b6099b-kube-api-access-84vh4\") pod \"7c521b14-3481-4326-9219-ec19d2b6099b\" (UID: \"7c521b14-3481-4326-9219-ec19d2b6099b\") " Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.996267 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content\") pod \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\" (UID: \"e25e5fd0-12f0-4b49-9cc6-aac032a3b270\") " Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.996796 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.996888 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d29f\" (UniqueName: \"kubernetes.io/projected/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-kube-api-access-7d29f\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:03 crc kubenswrapper[4810]: W1008 06:45:03.996930 4810 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/e25e5fd0-12f0-4b49-9cc6-aac032a3b270/volumes/kubernetes.io~empty-dir/catalog-content Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.997073 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e25e5fd0-12f0-4b49-9cc6-aac032a3b270" (UID: "e25e5fd0-12f0-4b49-9cc6-aac032a3b270"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:45:03 crc kubenswrapper[4810]: I1008 06:45:03.998217 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c521b14-3481-4326-9219-ec19d2b6099b-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c521b14-3481-4326-9219-ec19d2b6099b" (UID: "7c521b14-3481-4326-9219-ec19d2b6099b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.001000 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c521b14-3481-4326-9219-ec19d2b6099b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c521b14-3481-4326-9219-ec19d2b6099b" (UID: "7c521b14-3481-4326-9219-ec19d2b6099b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.000999 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c521b14-3481-4326-9219-ec19d2b6099b-kube-api-access-84vh4" (OuterVolumeSpecName: "kube-api-access-84vh4") pod "7c521b14-3481-4326-9219-ec19d2b6099b" (UID: "7c521b14-3481-4326-9219-ec19d2b6099b"). InnerVolumeSpecName "kube-api-access-84vh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.098258 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84vh4\" (UniqueName: \"kubernetes.io/projected/7c521b14-3481-4326-9219-ec19d2b6099b-kube-api-access-84vh4\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.098306 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e25e5fd0-12f0-4b49-9cc6-aac032a3b270-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.098320 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c521b14-3481-4326-9219-ec19d2b6099b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.098336 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c521b14-3481-4326-9219-ec19d2b6099b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.342189 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" event={"ID":"82648b77-ff5f-4c3b-93b9-3a860b9f4149","Type":"ContainerStarted","Data":"7bf69340103b888099a976390e91d132bfc7581f15351ed9352ac7c2b56827da"} Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.342244 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" event={"ID":"82648b77-ff5f-4c3b-93b9-3a860b9f4149","Type":"ContainerStarted","Data":"0cf64bf3f947d1eec413dbc6d59f72718678687aa211a677e11b3dba8cc1d5db"} Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.344042 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-7chth" event={"ID":"0d166b56-ad1e-44dc-96b3-58d1d3116642","Type":"ContainerStarted","Data":"14671f3b9170da18220e66655a0d8bb1b0677402c6ea62acfa5c1616d91871c4"} Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.345232 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.346984 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.346957 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331765-ttj67" event={"ID":"7c521b14-3481-4326-9219-ec19d2b6099b","Type":"ContainerDied","Data":"c7eac283dd6170ce048f1f698e5bd389a5c964978fe8ad47305cedc3cef00e6c"} Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.347141 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7eac283dd6170ce048f1f698e5bd389a5c964978fe8ad47305cedc3cef00e6c" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.350587 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbzq9" event={"ID":"e25e5fd0-12f0-4b49-9cc6-aac032a3b270","Type":"ContainerDied","Data":"77e18b42eae5edc87674b1d22720e549901628ab6b89dca080234974714b2280"} Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.350665 4810 scope.go:117] "RemoveContainer" containerID="bb1c6539055ec91f99bdf696636082bebc5ed1e6ee7f07a9cb3ca422559d80fd" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.351234 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbzq9" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.369770 4810 scope.go:117] "RemoveContainer" containerID="223dfe363327654ff5ec4a8b8cd0e476701c6406571a5c6393c2f0324ff9c1ce" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.371429 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-cfv7r" podStartSLOduration=2.371413074 podStartE2EDuration="2.371413074s" podCreationTimestamp="2025-10-08 06:45:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:45:04.368909266 +0000 UTC m=+807.003349016" watchObservedRunningTime="2025-10-08 06:45:04.371413074 +0000 UTC m=+807.005852814" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.390347 4810 scope.go:117] "RemoveContainer" containerID="066528a016f7b4f6d06b402231b6b44169060e34fea503879269c1f443245f56" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.420219 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-7chth" podStartSLOduration=2.45435757 podStartE2EDuration="7.420195701s" podCreationTimestamp="2025-10-08 06:44:57 +0000 UTC" firstStartedPulling="2025-10-08 06:44:58.314903506 +0000 UTC m=+800.949343246" lastFinishedPulling="2025-10-08 06:45:03.280741637 +0000 UTC m=+805.915181377" observedRunningTime="2025-10-08 06:45:04.403951676 +0000 UTC m=+807.038391436" watchObservedRunningTime="2025-10-08 06:45:04.420195701 +0000 UTC m=+807.054635441" Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.422840 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tbzq9"] Oct 08 06:45:04 crc kubenswrapper[4810]: I1008 06:45:04.428172 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tbzq9"] Oct 08 06:45:06 crc kubenswrapper[4810]: I1008 06:45:06.081709 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" path="/var/lib/kubelet/pods/e25e5fd0-12f0-4b49-9cc6-aac032a3b270/volumes" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.287577 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-ptlfz"] Oct 08 06:45:08 crc kubenswrapper[4810]: E1008 06:45:08.287894 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c521b14-3481-4326-9219-ec19d2b6099b" containerName="collect-profiles" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.287910 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c521b14-3481-4326-9219-ec19d2b6099b" containerName="collect-profiles" Oct 08 06:45:08 crc kubenswrapper[4810]: E1008 06:45:08.287926 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="extract-content" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.287936 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="extract-content" Oct 08 06:45:08 crc kubenswrapper[4810]: E1008 06:45:08.287953 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="registry-server" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.287981 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="registry-server" Oct 08 06:45:08 crc kubenswrapper[4810]: E1008 06:45:08.288002 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="extract-utilities" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.288010 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="extract-utilities" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.288130 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e25e5fd0-12f0-4b49-9cc6-aac032a3b270" containerName="registry-server" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.288148 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c521b14-3481-4326-9219-ec19d2b6099b" containerName="collect-profiles" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.288611 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.290782 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-g9tw7" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.303076 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-ptlfz"] Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.464026 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11ae85d1-fe3e-49ec-a984-1dc8e35050d6-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-ptlfz\" (UID: \"11ae85d1-fe3e-49ec-a984-1dc8e35050d6\") " pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.464625 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2gzl\" (UniqueName: \"kubernetes.io/projected/11ae85d1-fe3e-49ec-a984-1dc8e35050d6-kube-api-access-q2gzl\") pod \"cert-manager-7d4cc89fcb-ptlfz\" (UID: \"11ae85d1-fe3e-49ec-a984-1dc8e35050d6\") " pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.566448 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2gzl\" (UniqueName: \"kubernetes.io/projected/11ae85d1-fe3e-49ec-a984-1dc8e35050d6-kube-api-access-q2gzl\") pod \"cert-manager-7d4cc89fcb-ptlfz\" (UID: \"11ae85d1-fe3e-49ec-a984-1dc8e35050d6\") " pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.566586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11ae85d1-fe3e-49ec-a984-1dc8e35050d6-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-ptlfz\" (UID: \"11ae85d1-fe3e-49ec-a984-1dc8e35050d6\") " pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.597286 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2gzl\" (UniqueName: \"kubernetes.io/projected/11ae85d1-fe3e-49ec-a984-1dc8e35050d6-kube-api-access-q2gzl\") pod \"cert-manager-7d4cc89fcb-ptlfz\" (UID: \"11ae85d1-fe3e-49ec-a984-1dc8e35050d6\") " pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.597795 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11ae85d1-fe3e-49ec-a984-1dc8e35050d6-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-ptlfz\" (UID: \"11ae85d1-fe3e-49ec-a984-1dc8e35050d6\") " pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.612514 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" Oct 08 06:45:08 crc kubenswrapper[4810]: I1008 06:45:08.940853 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-ptlfz"] Oct 08 06:45:09 crc kubenswrapper[4810]: I1008 06:45:09.392124 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" event={"ID":"11ae85d1-fe3e-49ec-a984-1dc8e35050d6","Type":"ContainerStarted","Data":"000c649f3f8226b5468884e9919ec2c66e0acea370f14ee03d7eff53375f868e"} Oct 08 06:45:09 crc kubenswrapper[4810]: I1008 06:45:09.392606 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" event={"ID":"11ae85d1-fe3e-49ec-a984-1dc8e35050d6","Type":"ContainerStarted","Data":"5f24010364a26a914f7da98a8dee8fe6e1db7acaa1cb5176c9dc683cdc3a4942"} Oct 08 06:45:09 crc kubenswrapper[4810]: I1008 06:45:09.433631 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-ptlfz" podStartSLOduration=1.433591646 podStartE2EDuration="1.433591646s" podCreationTimestamp="2025-10-08 06:45:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:45:09.423465328 +0000 UTC m=+812.057905108" watchObservedRunningTime="2025-10-08 06:45:09.433591646 +0000 UTC m=+812.068031426" Oct 08 06:45:12 crc kubenswrapper[4810]: I1008 06:45:12.853524 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-7chth" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.272854 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-t7xgk"] Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.274157 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t7xgk" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.277857 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.278632 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.279348 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-kd2zl" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.304995 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t7xgk"] Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.401398 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9cv4\" (UniqueName: \"kubernetes.io/projected/4498b0f0-8c36-40e5-8cbc-caf980999945-kube-api-access-z9cv4\") pod \"openstack-operator-index-t7xgk\" (UID: \"4498b0f0-8c36-40e5-8cbc-caf980999945\") " pod="openstack-operators/openstack-operator-index-t7xgk" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.503099 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9cv4\" (UniqueName: \"kubernetes.io/projected/4498b0f0-8c36-40e5-8cbc-caf980999945-kube-api-access-z9cv4\") pod \"openstack-operator-index-t7xgk\" (UID: \"4498b0f0-8c36-40e5-8cbc-caf980999945\") " pod="openstack-operators/openstack-operator-index-t7xgk" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.528434 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9cv4\" (UniqueName: \"kubernetes.io/projected/4498b0f0-8c36-40e5-8cbc-caf980999945-kube-api-access-z9cv4\") pod \"openstack-operator-index-t7xgk\" (UID: \"4498b0f0-8c36-40e5-8cbc-caf980999945\") " pod="openstack-operators/openstack-operator-index-t7xgk" Oct 08 06:45:16 crc kubenswrapper[4810]: I1008 06:45:16.597835 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t7xgk" Oct 08 06:45:17 crc kubenswrapper[4810]: I1008 06:45:17.082706 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t7xgk"] Oct 08 06:45:17 crc kubenswrapper[4810]: I1008 06:45:17.467532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t7xgk" event={"ID":"4498b0f0-8c36-40e5-8cbc-caf980999945","Type":"ContainerStarted","Data":"864d17cea24171f96c89be21880af617f2f482f9d8fed94dac66ffb8f96b2f8c"} Oct 08 06:45:18 crc kubenswrapper[4810]: I1008 06:45:18.483331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t7xgk" event={"ID":"4498b0f0-8c36-40e5-8cbc-caf980999945","Type":"ContainerStarted","Data":"3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c"} Oct 08 06:45:18 crc kubenswrapper[4810]: I1008 06:45:18.506728 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-t7xgk" podStartSLOduration=1.54972835 podStartE2EDuration="2.506696413s" podCreationTimestamp="2025-10-08 06:45:16 +0000 UTC" firstStartedPulling="2025-10-08 06:45:17.094753019 +0000 UTC m=+819.729192769" lastFinishedPulling="2025-10-08 06:45:18.051721092 +0000 UTC m=+820.686160832" observedRunningTime="2025-10-08 06:45:18.505324886 +0000 UTC m=+821.139764666" watchObservedRunningTime="2025-10-08 06:45:18.506696413 +0000 UTC m=+821.141136193" Oct 08 06:45:19 crc kubenswrapper[4810]: I1008 06:45:19.629423 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-t7xgk"] Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.240363 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-q9z86"] Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.276181 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q9z86"] Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.276296 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.277141 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p92b\" (UniqueName: \"kubernetes.io/projected/402bba51-e1a6-4f79-823f-864c2d99a0e6-kube-api-access-5p92b\") pod \"openstack-operator-index-q9z86\" (UID: \"402bba51-e1a6-4f79-823f-864c2d99a0e6\") " pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.378677 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p92b\" (UniqueName: \"kubernetes.io/projected/402bba51-e1a6-4f79-823f-864c2d99a0e6-kube-api-access-5p92b\") pod \"openstack-operator-index-q9z86\" (UID: \"402bba51-e1a6-4f79-823f-864c2d99a0e6\") " pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.403980 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p92b\" (UniqueName: \"kubernetes.io/projected/402bba51-e1a6-4f79-823f-864c2d99a0e6-kube-api-access-5p92b\") pod \"openstack-operator-index-q9z86\" (UID: \"402bba51-e1a6-4f79-823f-864c2d99a0e6\") " pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.499652 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-t7xgk" podUID="4498b0f0-8c36-40e5-8cbc-caf980999945" containerName="registry-server" containerID="cri-o://3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c" gracePeriod=2 Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.596295 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:20 crc kubenswrapper[4810]: I1008 06:45:20.958599 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t7xgk" Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.088757 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9cv4\" (UniqueName: \"kubernetes.io/projected/4498b0f0-8c36-40e5-8cbc-caf980999945-kube-api-access-z9cv4\") pod \"4498b0f0-8c36-40e5-8cbc-caf980999945\" (UID: \"4498b0f0-8c36-40e5-8cbc-caf980999945\") " Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.094554 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4498b0f0-8c36-40e5-8cbc-caf980999945-kube-api-access-z9cv4" (OuterVolumeSpecName: "kube-api-access-z9cv4") pod "4498b0f0-8c36-40e5-8cbc-caf980999945" (UID: "4498b0f0-8c36-40e5-8cbc-caf980999945"). InnerVolumeSpecName "kube-api-access-z9cv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.127013 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q9z86"] Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.190680 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9cv4\" (UniqueName: \"kubernetes.io/projected/4498b0f0-8c36-40e5-8cbc-caf980999945-kube-api-access-z9cv4\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.510471 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q9z86" event={"ID":"402bba51-e1a6-4f79-823f-864c2d99a0e6","Type":"ContainerStarted","Data":"7365a8a246bf126125f402185d9b19e523817d076febe2d515ce0b92cc513584"} Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.513076 4810 generic.go:334] "Generic (PLEG): container finished" podID="4498b0f0-8c36-40e5-8cbc-caf980999945" containerID="3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c" exitCode=0 Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.513151 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t7xgk" event={"ID":"4498b0f0-8c36-40e5-8cbc-caf980999945","Type":"ContainerDied","Data":"3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c"} Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.513249 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t7xgk" event={"ID":"4498b0f0-8c36-40e5-8cbc-caf980999945","Type":"ContainerDied","Data":"864d17cea24171f96c89be21880af617f2f482f9d8fed94dac66ffb8f96b2f8c"} Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.513289 4810 scope.go:117] "RemoveContainer" containerID="3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c" Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.513284 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t7xgk" Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.550521 4810 scope.go:117] "RemoveContainer" containerID="3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c" Oct 08 06:45:21 crc kubenswrapper[4810]: E1008 06:45:21.552018 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c\": container with ID starting with 3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c not found: ID does not exist" containerID="3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c" Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.552099 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c"} err="failed to get container status \"3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c\": rpc error: code = NotFound desc = could not find container \"3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c\": container with ID starting with 3375fb87a3719ea8f55d8a47e97d1891fc80a32234f02d4d4c857ab72e73828c not found: ID does not exist" Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.565176 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-t7xgk"] Oct 08 06:45:21 crc kubenswrapper[4810]: I1008 06:45:21.571154 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-t7xgk"] Oct 08 06:45:22 crc kubenswrapper[4810]: I1008 06:45:22.092304 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4498b0f0-8c36-40e5-8cbc-caf980999945" path="/var/lib/kubelet/pods/4498b0f0-8c36-40e5-8cbc-caf980999945/volumes" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.545200 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q9z86" event={"ID":"402bba51-e1a6-4f79-823f-864c2d99a0e6","Type":"ContainerStarted","Data":"19ced22aecb8873a79ed7f2850e5b66efc5eddc716fc1f869e2bdad200ecc655"} Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.575329 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-q9z86" podStartSLOduration=1.909955681 podStartE2EDuration="4.575294713s" podCreationTimestamp="2025-10-08 06:45:20 +0000 UTC" firstStartedPulling="2025-10-08 06:45:21.142903475 +0000 UTC m=+823.777343225" lastFinishedPulling="2025-10-08 06:45:23.808242497 +0000 UTC m=+826.442682257" observedRunningTime="2025-10-08 06:45:24.569609657 +0000 UTC m=+827.204049437" watchObservedRunningTime="2025-10-08 06:45:24.575294713 +0000 UTC m=+827.209734483" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.646611 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9v25c"] Oct 08 06:45:24 crc kubenswrapper[4810]: E1008 06:45:24.647019 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4498b0f0-8c36-40e5-8cbc-caf980999945" containerName="registry-server" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.647036 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4498b0f0-8c36-40e5-8cbc-caf980999945" containerName="registry-server" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.647209 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4498b0f0-8c36-40e5-8cbc-caf980999945" containerName="registry-server" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.648552 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.650324 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-utilities\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.650391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-catalog-content\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.650439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkw74\" (UniqueName: \"kubernetes.io/projected/2370327d-3f0e-444a-bb1c-8e50d5061f16-kube-api-access-gkw74\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.668087 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9v25c"] Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.751898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-utilities\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.751998 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-catalog-content\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.752043 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkw74\" (UniqueName: \"kubernetes.io/projected/2370327d-3f0e-444a-bb1c-8e50d5061f16-kube-api-access-gkw74\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.752840 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-utilities\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.752921 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-catalog-content\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.785180 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkw74\" (UniqueName: \"kubernetes.io/projected/2370327d-3f0e-444a-bb1c-8e50d5061f16-kube-api-access-gkw74\") pod \"certified-operators-9v25c\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:24 crc kubenswrapper[4810]: I1008 06:45:24.985116 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:25 crc kubenswrapper[4810]: I1008 06:45:25.504846 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9v25c"] Oct 08 06:45:25 crc kubenswrapper[4810]: W1008 06:45:25.511739 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2370327d_3f0e_444a_bb1c_8e50d5061f16.slice/crio-be3eda4ef1f99dca467c6a2d32e570017bc1af41bf17a31daea47c6d402138bb WatchSource:0}: Error finding container be3eda4ef1f99dca467c6a2d32e570017bc1af41bf17a31daea47c6d402138bb: Status 404 returned error can't find the container with id be3eda4ef1f99dca467c6a2d32e570017bc1af41bf17a31daea47c6d402138bb Oct 08 06:45:25 crc kubenswrapper[4810]: I1008 06:45:25.561507 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9v25c" event={"ID":"2370327d-3f0e-444a-bb1c-8e50d5061f16","Type":"ContainerStarted","Data":"be3eda4ef1f99dca467c6a2d32e570017bc1af41bf17a31daea47c6d402138bb"} Oct 08 06:45:26 crc kubenswrapper[4810]: I1008 06:45:26.575110 4810 generic.go:334] "Generic (PLEG): container finished" podID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerID="372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484" exitCode=0 Oct 08 06:45:26 crc kubenswrapper[4810]: I1008 06:45:26.575196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9v25c" event={"ID":"2370327d-3f0e-444a-bb1c-8e50d5061f16","Type":"ContainerDied","Data":"372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484"} Oct 08 06:45:27 crc kubenswrapper[4810]: I1008 06:45:27.587622 4810 generic.go:334] "Generic (PLEG): container finished" podID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerID="da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca" exitCode=0 Oct 08 06:45:27 crc kubenswrapper[4810]: I1008 06:45:27.587806 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9v25c" event={"ID":"2370327d-3f0e-444a-bb1c-8e50d5061f16","Type":"ContainerDied","Data":"da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca"} Oct 08 06:45:28 crc kubenswrapper[4810]: I1008 06:45:28.599882 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9v25c" event={"ID":"2370327d-3f0e-444a-bb1c-8e50d5061f16","Type":"ContainerStarted","Data":"6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185"} Oct 08 06:45:28 crc kubenswrapper[4810]: I1008 06:45:28.638369 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9v25c" podStartSLOduration=3.126165805 podStartE2EDuration="4.638334496s" podCreationTimestamp="2025-10-08 06:45:24 +0000 UTC" firstStartedPulling="2025-10-08 06:45:26.580385535 +0000 UTC m=+829.214825315" lastFinishedPulling="2025-10-08 06:45:28.092554256 +0000 UTC m=+830.726994006" observedRunningTime="2025-10-08 06:45:28.627269023 +0000 UTC m=+831.261708803" watchObservedRunningTime="2025-10-08 06:45:28.638334496 +0000 UTC m=+831.272774276" Oct 08 06:45:30 crc kubenswrapper[4810]: I1008 06:45:30.596487 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:30 crc kubenswrapper[4810]: I1008 06:45:30.597047 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:30 crc kubenswrapper[4810]: I1008 06:45:30.648273 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:30 crc kubenswrapper[4810]: I1008 06:45:30.701647 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-q9z86" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.704113 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb"] Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.706842 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.710024 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-5vcp4" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.715218 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb"] Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.794582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-util\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.794767 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-bundle\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.794803 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmbsk\" (UniqueName: \"kubernetes.io/projected/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-kube-api-access-jmbsk\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.895645 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-bundle\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.895740 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmbsk\" (UniqueName: \"kubernetes.io/projected/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-kube-api-access-jmbsk\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.895847 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-util\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.896385 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-bundle\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.896444 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-util\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:31 crc kubenswrapper[4810]: I1008 06:45:31.918570 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmbsk\" (UniqueName: \"kubernetes.io/projected/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-kube-api-access-jmbsk\") pod \"36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:32 crc kubenswrapper[4810]: I1008 06:45:32.033980 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:32 crc kubenswrapper[4810]: I1008 06:45:32.405385 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb"] Oct 08 06:45:32 crc kubenswrapper[4810]: I1008 06:45:32.631408 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" event={"ID":"bd9febeb-3b39-4d90-9c95-76f86ac8b87e","Type":"ContainerStarted","Data":"b075afe5f8743fc020e2c12c22ba471ea8c0b566708bba85aa89bde2a78dcfdd"} Oct 08 06:45:32 crc kubenswrapper[4810]: I1008 06:45:32.631464 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" event={"ID":"bd9febeb-3b39-4d90-9c95-76f86ac8b87e","Type":"ContainerStarted","Data":"707ba079500a8c5c21147f8678dd4382c37dbff1123648c74175454e491652da"} Oct 08 06:45:33 crc kubenswrapper[4810]: I1008 06:45:33.638393 4810 generic.go:334] "Generic (PLEG): container finished" podID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerID="b075afe5f8743fc020e2c12c22ba471ea8c0b566708bba85aa89bde2a78dcfdd" exitCode=0 Oct 08 06:45:33 crc kubenswrapper[4810]: I1008 06:45:33.638446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" event={"ID":"bd9febeb-3b39-4d90-9c95-76f86ac8b87e","Type":"ContainerDied","Data":"b075afe5f8743fc020e2c12c22ba471ea8c0b566708bba85aa89bde2a78dcfdd"} Oct 08 06:45:34 crc kubenswrapper[4810]: I1008 06:45:34.985997 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:34 crc kubenswrapper[4810]: I1008 06:45:34.986484 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:35 crc kubenswrapper[4810]: I1008 06:45:35.047154 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:35 crc kubenswrapper[4810]: I1008 06:45:35.664445 4810 generic.go:334] "Generic (PLEG): container finished" podID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerID="fe06ab92bd27a388634beac8c3fc383e94b75e34c9f40bbced7f27f06b4eae84" exitCode=0 Oct 08 06:45:35 crc kubenswrapper[4810]: I1008 06:45:35.664709 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" event={"ID":"bd9febeb-3b39-4d90-9c95-76f86ac8b87e","Type":"ContainerDied","Data":"fe06ab92bd27a388634beac8c3fc383e94b75e34c9f40bbced7f27f06b4eae84"} Oct 08 06:45:35 crc kubenswrapper[4810]: I1008 06:45:35.740321 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:36 crc kubenswrapper[4810]: I1008 06:45:36.680763 4810 generic.go:334] "Generic (PLEG): container finished" podID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerID="183fc19dc3e56734f669f6fd7f060b46274d123454a9cd2f0c344d2b23fec2d3" exitCode=0 Oct 08 06:45:36 crc kubenswrapper[4810]: I1008 06:45:36.680841 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" event={"ID":"bd9febeb-3b39-4d90-9c95-76f86ac8b87e","Type":"ContainerDied","Data":"183fc19dc3e56734f669f6fd7f060b46274d123454a9cd2f0c344d2b23fec2d3"} Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.045528 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9v25c"] Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.046062 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9v25c" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="registry-server" containerID="cri-o://6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185" gracePeriod=2 Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.080948 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.232147 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-util\") pod \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.232205 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmbsk\" (UniqueName: \"kubernetes.io/projected/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-kube-api-access-jmbsk\") pod \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.232302 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-bundle\") pod \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\" (UID: \"bd9febeb-3b39-4d90-9c95-76f86ac8b87e\") " Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.234378 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-bundle" (OuterVolumeSpecName: "bundle") pod "bd9febeb-3b39-4d90-9c95-76f86ac8b87e" (UID: "bd9febeb-3b39-4d90-9c95-76f86ac8b87e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.238466 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-kube-api-access-jmbsk" (OuterVolumeSpecName: "kube-api-access-jmbsk") pod "bd9febeb-3b39-4d90-9c95-76f86ac8b87e" (UID: "bd9febeb-3b39-4d90-9c95-76f86ac8b87e"). InnerVolumeSpecName "kube-api-access-jmbsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.333521 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmbsk\" (UniqueName: \"kubernetes.io/projected/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-kube-api-access-jmbsk\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.333926 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.356648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-util" (OuterVolumeSpecName: "util") pod "bd9febeb-3b39-4d90-9c95-76f86ac8b87e" (UID: "bd9febeb-3b39-4d90-9c95-76f86ac8b87e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.435931 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd9febeb-3b39-4d90-9c95-76f86ac8b87e-util\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.489923 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.643807 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-utilities\") pod \"2370327d-3f0e-444a-bb1c-8e50d5061f16\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.644014 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkw74\" (UniqueName: \"kubernetes.io/projected/2370327d-3f0e-444a-bb1c-8e50d5061f16-kube-api-access-gkw74\") pod \"2370327d-3f0e-444a-bb1c-8e50d5061f16\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.644100 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-catalog-content\") pod \"2370327d-3f0e-444a-bb1c-8e50d5061f16\" (UID: \"2370327d-3f0e-444a-bb1c-8e50d5061f16\") " Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.645043 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-utilities" (OuterVolumeSpecName: "utilities") pod "2370327d-3f0e-444a-bb1c-8e50d5061f16" (UID: "2370327d-3f0e-444a-bb1c-8e50d5061f16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.648070 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2370327d-3f0e-444a-bb1c-8e50d5061f16-kube-api-access-gkw74" (OuterVolumeSpecName: "kube-api-access-gkw74") pod "2370327d-3f0e-444a-bb1c-8e50d5061f16" (UID: "2370327d-3f0e-444a-bb1c-8e50d5061f16"). InnerVolumeSpecName "kube-api-access-gkw74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.698011 4810 generic.go:334] "Generic (PLEG): container finished" podID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerID="6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185" exitCode=0 Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.698075 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9v25c" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.698100 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9v25c" event={"ID":"2370327d-3f0e-444a-bb1c-8e50d5061f16","Type":"ContainerDied","Data":"6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185"} Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.698172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9v25c" event={"ID":"2370327d-3f0e-444a-bb1c-8e50d5061f16","Type":"ContainerDied","Data":"be3eda4ef1f99dca467c6a2d32e570017bc1af41bf17a31daea47c6d402138bb"} Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.698197 4810 scope.go:117] "RemoveContainer" containerID="6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.702987 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" event={"ID":"bd9febeb-3b39-4d90-9c95-76f86ac8b87e","Type":"ContainerDied","Data":"707ba079500a8c5c21147f8678dd4382c37dbff1123648c74175454e491652da"} Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.703030 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="707ba079500a8c5c21147f8678dd4382c37dbff1123648c74175454e491652da" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.703102 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.713422 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2370327d-3f0e-444a-bb1c-8e50d5061f16" (UID: "2370327d-3f0e-444a-bb1c-8e50d5061f16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.723078 4810 scope.go:117] "RemoveContainer" containerID="da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.746588 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkw74\" (UniqueName: \"kubernetes.io/projected/2370327d-3f0e-444a-bb1c-8e50d5061f16-kube-api-access-gkw74\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.746640 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.746656 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2370327d-3f0e-444a-bb1c-8e50d5061f16-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.753240 4810 scope.go:117] "RemoveContainer" containerID="372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.771521 4810 scope.go:117] "RemoveContainer" containerID="6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185" Oct 08 06:45:38 crc kubenswrapper[4810]: E1008 06:45:38.772199 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185\": container with ID starting with 6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185 not found: ID does not exist" containerID="6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.772255 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185"} err="failed to get container status \"6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185\": rpc error: code = NotFound desc = could not find container \"6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185\": container with ID starting with 6374f6a42851be58e8b29fcc4edd71185f7446279373d9f1412ac4d671581185 not found: ID does not exist" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.772294 4810 scope.go:117] "RemoveContainer" containerID="da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca" Oct 08 06:45:38 crc kubenswrapper[4810]: E1008 06:45:38.772705 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca\": container with ID starting with da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca not found: ID does not exist" containerID="da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.772753 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca"} err="failed to get container status \"da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca\": rpc error: code = NotFound desc = could not find container \"da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca\": container with ID starting with da44c3cf86a60a63822ef60a1ebfc3f277300168cb09d1cd6839bd0fe3c24aca not found: ID does not exist" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.772791 4810 scope.go:117] "RemoveContainer" containerID="372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484" Oct 08 06:45:38 crc kubenswrapper[4810]: E1008 06:45:38.773068 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484\": container with ID starting with 372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484 not found: ID does not exist" containerID="372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484" Oct 08 06:45:38 crc kubenswrapper[4810]: I1008 06:45:38.773103 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484"} err="failed to get container status \"372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484\": rpc error: code = NotFound desc = could not find container \"372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484\": container with ID starting with 372d1ca4c9abb190df23c4d70a62e2efa6a11846d8bf007f957d6a4b1d429484 not found: ID does not exist" Oct 08 06:45:39 crc kubenswrapper[4810]: I1008 06:45:39.047469 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9v25c"] Oct 08 06:45:39 crc kubenswrapper[4810]: I1008 06:45:39.056742 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9v25c"] Oct 08 06:45:40 crc kubenswrapper[4810]: I1008 06:45:40.084130 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" path="/var/lib/kubelet/pods/2370327d-3f0e-444a-bb1c-8e50d5061f16/volumes" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.196658 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl"] Oct 08 06:45:42 crc kubenswrapper[4810]: E1008 06:45:42.198066 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerName="extract" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.198160 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerName="extract" Oct 08 06:45:42 crc kubenswrapper[4810]: E1008 06:45:42.198238 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="extract-utilities" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.198299 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="extract-utilities" Oct 08 06:45:42 crc kubenswrapper[4810]: E1008 06:45:42.198362 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerName="util" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.198444 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerName="util" Oct 08 06:45:42 crc kubenswrapper[4810]: E1008 06:45:42.198504 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="extract-content" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.198559 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="extract-content" Oct 08 06:45:42 crc kubenswrapper[4810]: E1008 06:45:42.198620 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="registry-server" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.198673 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="registry-server" Oct 08 06:45:42 crc kubenswrapper[4810]: E1008 06:45:42.198746 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerName="pull" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.198803 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerName="pull" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.198982 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd9febeb-3b39-4d90-9c95-76f86ac8b87e" containerName="extract" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.199071 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2370327d-3f0e-444a-bb1c-8e50d5061f16" containerName="registry-server" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.199776 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.209985 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-sqhqj" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.267603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl"] Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.302991 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2p7h\" (UniqueName: \"kubernetes.io/projected/71fb758c-689c-4098-b73a-e45b4bf6db60-kube-api-access-b2p7h\") pod \"openstack-operator-controller-operator-b6d857f89-wtlfl\" (UID: \"71fb758c-689c-4098-b73a-e45b4bf6db60\") " pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.404487 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2p7h\" (UniqueName: \"kubernetes.io/projected/71fb758c-689c-4098-b73a-e45b4bf6db60-kube-api-access-b2p7h\") pod \"openstack-operator-controller-operator-b6d857f89-wtlfl\" (UID: \"71fb758c-689c-4098-b73a-e45b4bf6db60\") " pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.430223 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2p7h\" (UniqueName: \"kubernetes.io/projected/71fb758c-689c-4098-b73a-e45b4bf6db60-kube-api-access-b2p7h\") pod \"openstack-operator-controller-operator-b6d857f89-wtlfl\" (UID: \"71fb758c-689c-4098-b73a-e45b4bf6db60\") " pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.517358 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" Oct 08 06:45:42 crc kubenswrapper[4810]: I1008 06:45:42.996130 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl"] Oct 08 06:45:43 crc kubenswrapper[4810]: I1008 06:45:43.770705 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" event={"ID":"71fb758c-689c-4098-b73a-e45b4bf6db60","Type":"ContainerStarted","Data":"80b2120a7bca9a571f08c27ba86e8a5cb11e5e6c67452ba93c7fb20f166df613"} Oct 08 06:45:48 crc kubenswrapper[4810]: I1008 06:45:48.841010 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" event={"ID":"71fb758c-689c-4098-b73a-e45b4bf6db60","Type":"ContainerStarted","Data":"2dfdf207748cb8b46e1d45a9e7362c67fe03cf19d3ae960c94f9e63d34e5f06e"} Oct 08 06:45:50 crc kubenswrapper[4810]: I1008 06:45:50.862212 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" event={"ID":"71fb758c-689c-4098-b73a-e45b4bf6db60","Type":"ContainerStarted","Data":"a88b2a69d930c90bedf7badfd10ca1295815ebc37508477d47bd8bbe99f38652"} Oct 08 06:45:50 crc kubenswrapper[4810]: I1008 06:45:50.862739 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" Oct 08 06:45:50 crc kubenswrapper[4810]: I1008 06:45:50.908271 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" podStartSLOduration=1.63032626 podStartE2EDuration="8.908243508s" podCreationTimestamp="2025-10-08 06:45:42 +0000 UTC" firstStartedPulling="2025-10-08 06:45:43.00870224 +0000 UTC m=+845.643142000" lastFinishedPulling="2025-10-08 06:45:50.286619508 +0000 UTC m=+852.921059248" observedRunningTime="2025-10-08 06:45:50.904831794 +0000 UTC m=+853.539271534" watchObservedRunningTime="2025-10-08 06:45:50.908243508 +0000 UTC m=+853.542683248" Oct 08 06:46:02 crc kubenswrapper[4810]: I1008 06:46:02.521414 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-b6d857f89-wtlfl" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.505328 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.509583 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.511949 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.516940 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.517259 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-hccf5" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.519735 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.521127 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-88vt8" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.527237 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.533627 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2jh2d" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.538040 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.543932 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.561605 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.562876 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.566544 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-r8j5n" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.568455 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.573258 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.574567 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.582349 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.582400 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-vwtk5" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.595210 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.595871 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgg92\" (UniqueName: \"kubernetes.io/projected/3fe805b1-a990-4683-8549-f73ebe9d850c-kube-api-access-rgg92\") pod \"barbican-operator-controller-manager-64f56ff694-tlndl\" (UID: \"3fe805b1-a990-4683-8549-f73ebe9d850c\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.596608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh755\" (UniqueName: \"kubernetes.io/projected/2c5ab097-4362-44a0-8f0a-a2a2483851f4-kube-api-access-mh755\") pod \"cinder-operator-controller-manager-84bd8f6848-2lq4z\" (UID: \"2c5ab097-4362-44a0-8f0a-a2a2483851f4\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.596648 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqt9z\" (UniqueName: \"kubernetes.io/projected/75590e1d-5795-4ebe-afd0-9abb818efd63-kube-api-access-gqt9z\") pod \"designate-operator-controller-manager-58d86cd59d-h4wk6\" (UID: \"75590e1d-5795-4ebe-afd0-9abb818efd63\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.625489 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.627436 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.630102 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-2pxt9" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.630284 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.663254 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.679594 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4js7b" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.688034 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.708927 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.716878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqt9z\" (UniqueName: \"kubernetes.io/projected/75590e1d-5795-4ebe-afd0-9abb818efd63-kube-api-access-gqt9z\") pod \"designate-operator-controller-manager-58d86cd59d-h4wk6\" (UID: \"75590e1d-5795-4ebe-afd0-9abb818efd63\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.716933 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tvv2\" (UniqueName: \"kubernetes.io/projected/6fc03514-36ec-4d41-8ad1-306ff2767a2f-kube-api-access-6tvv2\") pod \"heat-operator-controller-manager-7ccfc8cf49-ntp7k\" (UID: \"6fc03514-36ec-4d41-8ad1-306ff2767a2f\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.719209 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gv5q\" (UniqueName: \"kubernetes.io/projected/72a051e1-5079-45b0-b3d8-0592f872c85a-kube-api-access-5gv5q\") pod \"horizon-operator-controller-manager-5b477879bc-8jhkm\" (UID: \"72a051e1-5079-45b0-b3d8-0592f872c85a\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.719259 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nq4c\" (UniqueName: \"kubernetes.io/projected/0bebe3c8-1035-40b4-b533-4c214423f6b2-kube-api-access-5nq4c\") pod \"glance-operator-controller-manager-fd648f65-nwwg7\" (UID: \"0bebe3c8-1035-40b4-b533-4c214423f6b2\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.719298 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgg92\" (UniqueName: \"kubernetes.io/projected/3fe805b1-a990-4683-8549-f73ebe9d850c-kube-api-access-rgg92\") pod \"barbican-operator-controller-manager-64f56ff694-tlndl\" (UID: \"3fe805b1-a990-4683-8549-f73ebe9d850c\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.719330 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh755\" (UniqueName: \"kubernetes.io/projected/2c5ab097-4362-44a0-8f0a-a2a2483851f4-kube-api-access-mh755\") pod \"cinder-operator-controller-manager-84bd8f6848-2lq4z\" (UID: \"2c5ab097-4362-44a0-8f0a-a2a2483851f4\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.731505 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.744483 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.745791 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.751767 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgg92\" (UniqueName: \"kubernetes.io/projected/3fe805b1-a990-4683-8549-f73ebe9d850c-kube-api-access-rgg92\") pod \"barbican-operator-controller-manager-64f56ff694-tlndl\" (UID: \"3fe805b1-a990-4683-8549-f73ebe9d850c\") " pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.755541 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-nh5jn" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.756310 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh755\" (UniqueName: \"kubernetes.io/projected/2c5ab097-4362-44a0-8f0a-a2a2483851f4-kube-api-access-mh755\") pod \"cinder-operator-controller-manager-84bd8f6848-2lq4z\" (UID: \"2c5ab097-4362-44a0-8f0a-a2a2483851f4\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.764075 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.765472 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.777306 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-xx9st" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.780664 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqt9z\" (UniqueName: \"kubernetes.io/projected/75590e1d-5795-4ebe-afd0-9abb818efd63-kube-api-access-gqt9z\") pod \"designate-operator-controller-manager-58d86cd59d-h4wk6\" (UID: \"75590e1d-5795-4ebe-afd0-9abb818efd63\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.796690 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.818418 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.819952 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.828523 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nq4c\" (UniqueName: \"kubernetes.io/projected/0bebe3c8-1035-40b4-b533-4c214423f6b2-kube-api-access-5nq4c\") pod \"glance-operator-controller-manager-fd648f65-nwwg7\" (UID: \"0bebe3c8-1035-40b4-b533-4c214423f6b2\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.828759 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dcjb\" (UniqueName: \"kubernetes.io/projected/8181e145-9c88-47f0-a9a8-dc03947b75ef-kube-api-access-7dcjb\") pod \"ironic-operator-controller-manager-5467f8988c-s4m9g\" (UID: \"8181e145-9c88-47f0-a9a8-dc03947b75ef\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.828808 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tvv2\" (UniqueName: \"kubernetes.io/projected/6fc03514-36ec-4d41-8ad1-306ff2767a2f-kube-api-access-6tvv2\") pod \"heat-operator-controller-manager-7ccfc8cf49-ntp7k\" (UID: \"6fc03514-36ec-4d41-8ad1-306ff2767a2f\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.828867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-css7w\" (UniqueName: \"kubernetes.io/projected/0b266b2d-a60b-4990-b720-800d655a4b5f-kube-api-access-css7w\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.829048 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gv5q\" (UniqueName: \"kubernetes.io/projected/72a051e1-5079-45b0-b3d8-0592f872c85a-kube-api-access-5gv5q\") pod \"horizon-operator-controller-manager-5b477879bc-8jhkm\" (UID: \"72a051e1-5079-45b0-b3d8-0592f872c85a\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.829096 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n55c\" (UniqueName: \"kubernetes.io/projected/24b64e0e-8eb4-4235-9086-8f1f61f2bd8e-kube-api-access-7n55c\") pod \"keystone-operator-controller-manager-5b84cc7657-9gdl9\" (UID: \"24b64e0e-8eb4-4235-9086-8f1f61f2bd8e\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.829119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.830002 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8gr9t" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.833821 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.847916 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.854293 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nq4c\" (UniqueName: \"kubernetes.io/projected/0bebe3c8-1035-40b4-b533-4c214423f6b2-kube-api-access-5nq4c\") pod \"glance-operator-controller-manager-fd648f65-nwwg7\" (UID: \"0bebe3c8-1035-40b4-b533-4c214423f6b2\") " pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.859833 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.861319 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.874626 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tvv2\" (UniqueName: \"kubernetes.io/projected/6fc03514-36ec-4d41-8ad1-306ff2767a2f-kube-api-access-6tvv2\") pod \"heat-operator-controller-manager-7ccfc8cf49-ntp7k\" (UID: \"6fc03514-36ec-4d41-8ad1-306ff2767a2f\") " pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.877932 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gv5q\" (UniqueName: \"kubernetes.io/projected/72a051e1-5079-45b0-b3d8-0592f872c85a-kube-api-access-5gv5q\") pod \"horizon-operator-controller-manager-5b477879bc-8jhkm\" (UID: \"72a051e1-5079-45b0-b3d8-0592f872c85a\") " pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.879504 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.885316 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.894120 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.900667 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-6s2vs" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.914535 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.917103 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.929179 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.930123 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dcjb\" (UniqueName: \"kubernetes.io/projected/8181e145-9c88-47f0-a9a8-dc03947b75ef-kube-api-access-7dcjb\") pod \"ironic-operator-controller-manager-5467f8988c-s4m9g\" (UID: \"8181e145-9c88-47f0-a9a8-dc03947b75ef\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.930186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-css7w\" (UniqueName: \"kubernetes.io/projected/0b266b2d-a60b-4990-b720-800d655a4b5f-kube-api-access-css7w\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.930247 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.930267 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n55c\" (UniqueName: \"kubernetes.io/projected/24b64e0e-8eb4-4235-9086-8f1f61f2bd8e-kube-api-access-7n55c\") pod \"keystone-operator-controller-manager-5b84cc7657-9gdl9\" (UID: \"24b64e0e-8eb4-4235-9086-8f1f61f2bd8e\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.930306 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkfrc\" (UniqueName: \"kubernetes.io/projected/e8deb064-ce04-4e12-ab87-e4853e499c20-kube-api-access-wkfrc\") pod \"manila-operator-controller-manager-7cb48dbc-657k2\" (UID: \"e8deb064-ce04-4e12-ab87-e4853e499c20\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" Oct 08 06:46:18 crc kubenswrapper[4810]: E1008 06:46:18.930712 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 06:46:18 crc kubenswrapper[4810]: E1008 06:46:18.930785 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert podName:0b266b2d-a60b-4990-b720-800d655a4b5f nodeName:}" failed. No retries permitted until 2025-10-08 06:46:19.430765478 +0000 UTC m=+882.065205218 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert") pod "infra-operator-controller-manager-84788b6bc5-dqjfx" (UID: "0b266b2d-a60b-4990-b720-800d655a4b5f") : secret "infra-operator-webhook-server-cert" not found Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.953726 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n55c\" (UniqueName: \"kubernetes.io/projected/24b64e0e-8eb4-4235-9086-8f1f61f2bd8e-kube-api-access-7n55c\") pod \"keystone-operator-controller-manager-5b84cc7657-9gdl9\" (UID: \"24b64e0e-8eb4-4235-9086-8f1f61f2bd8e\") " pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.954466 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dcjb\" (UniqueName: \"kubernetes.io/projected/8181e145-9c88-47f0-a9a8-dc03947b75ef-kube-api-access-7dcjb\") pod \"ironic-operator-controller-manager-5467f8988c-s4m9g\" (UID: \"8181e145-9c88-47f0-a9a8-dc03947b75ef\") " pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.957582 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-css7w\" (UniqueName: \"kubernetes.io/projected/0b266b2d-a60b-4990-b720-800d655a4b5f-kube-api-access-css7w\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.964778 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.965909 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.968597 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-dgdx8" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.974930 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf"] Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.979191 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" Oct 08 06:46:18 crc kubenswrapper[4810]: I1008 06:46:18.985509 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-zpdt4" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.013596 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.015713 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.016826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.020889 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-pn6x8" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.026995 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.032869 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkfrc\" (UniqueName: \"kubernetes.io/projected/e8deb064-ce04-4e12-ab87-e4853e499c20-kube-api-access-wkfrc\") pod \"manila-operator-controller-manager-7cb48dbc-657k2\" (UID: \"e8deb064-ce04-4e12-ab87-e4853e499c20\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.033487 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzn64\" (UniqueName: \"kubernetes.io/projected/c19839d0-64b9-4f39-a4e2-40cdc62bd9b1-kube-api-access-xzn64\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-ckvcg\" (UID: \"c19839d0-64b9-4f39-a4e2-40cdc62bd9b1\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.033577 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdr69\" (UniqueName: \"kubernetes.io/projected/681220bd-766a-4d1e-970b-662b4c64aaf6-kube-api-access-bdr69\") pod \"nova-operator-controller-manager-6c9b57c67-8cmwf\" (UID: \"681220bd-766a-4d1e-970b-662b4c64aaf6\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.033645 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksbhd\" (UniqueName: \"kubernetes.io/projected/09c37aa7-fb5d-4984-9f32-86a4a323ea49-kube-api-access-ksbhd\") pod \"neutron-operator-controller-manager-69b956fbf6-r9g7m\" (UID: \"09c37aa7-fb5d-4984-9f32-86a4a323ea49\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.033731 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.063881 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkfrc\" (UniqueName: \"kubernetes.io/projected/e8deb064-ce04-4e12-ab87-e4853e499c20-kube-api-access-wkfrc\") pod \"manila-operator-controller-manager-7cb48dbc-657k2\" (UID: \"e8deb064-ce04-4e12-ab87-e4853e499c20\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.076801 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.092318 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.094329 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.098645 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.103687 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xs4xw" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.106758 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.109140 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.111157 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nqlkb" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.115648 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.134816 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzn64\" (UniqueName: \"kubernetes.io/projected/c19839d0-64b9-4f39-a4e2-40cdc62bd9b1-kube-api-access-xzn64\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-ckvcg\" (UID: \"c19839d0-64b9-4f39-a4e2-40cdc62bd9b1\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.134859 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmw7z\" (UniqueName: \"kubernetes.io/projected/413157c3-1aa7-425d-98ae-e44a7c19b425-kube-api-access-vmw7z\") pod \"octavia-operator-controller-manager-69f59f9d8-rtmss\" (UID: \"413157c3-1aa7-425d-98ae-e44a7c19b425\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.134884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdr69\" (UniqueName: \"kubernetes.io/projected/681220bd-766a-4d1e-970b-662b4c64aaf6-kube-api-access-bdr69\") pod \"nova-operator-controller-manager-6c9b57c67-8cmwf\" (UID: \"681220bd-766a-4d1e-970b-662b4c64aaf6\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.134922 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksbhd\" (UniqueName: \"kubernetes.io/projected/09c37aa7-fb5d-4984-9f32-86a4a323ea49-kube-api-access-ksbhd\") pod \"neutron-operator-controller-manager-69b956fbf6-r9g7m\" (UID: \"09c37aa7-fb5d-4984-9f32-86a4a323ea49\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.136111 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.137111 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.137319 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.146495 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-rkfp8" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.154390 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-qfp99"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.155498 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.168922 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-qjcx2" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.174566 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.183163 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdr69\" (UniqueName: \"kubernetes.io/projected/681220bd-766a-4d1e-970b-662b4c64aaf6-kube-api-access-bdr69\") pod \"nova-operator-controller-manager-6c9b57c67-8cmwf\" (UID: \"681220bd-766a-4d1e-970b-662b4c64aaf6\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.183700 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksbhd\" (UniqueName: \"kubernetes.io/projected/09c37aa7-fb5d-4984-9f32-86a4a323ea49-kube-api-access-ksbhd\") pod \"neutron-operator-controller-manager-69b956fbf6-r9g7m\" (UID: \"09c37aa7-fb5d-4984-9f32-86a4a323ea49\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.187017 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzn64\" (UniqueName: \"kubernetes.io/projected/c19839d0-64b9-4f39-a4e2-40cdc62bd9b1-kube-api-access-xzn64\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-ckvcg\" (UID: \"c19839d0-64b9-4f39-a4e2-40cdc62bd9b1\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.259674 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.260114 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmw7z\" (UniqueName: \"kubernetes.io/projected/413157c3-1aa7-425d-98ae-e44a7c19b425-kube-api-access-vmw7z\") pod \"octavia-operator-controller-manager-69f59f9d8-rtmss\" (UID: \"413157c3-1aa7-425d-98ae-e44a7c19b425\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.260247 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.260865 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.260150 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74z48\" (UniqueName: \"kubernetes.io/projected/cb1fd16d-a191-4086-9a18-37b74183042d-kube-api-access-74z48\") pod \"ovn-operator-controller-manager-54d485fd9-29jxc\" (UID: \"cb1fd16d-a191-4086-9a18-37b74183042d\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.283894 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.284619 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpm6f\" (UniqueName: \"kubernetes.io/projected/f164aece-1b13-46b7-9da2-16967eb942f5-kube-api-access-tpm6f\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.284721 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrtc5\" (UniqueName: \"kubernetes.io/projected/914ce43c-e528-4921-b1b4-aa47cdf4b004-kube-api-access-hrtc5\") pod \"swift-operator-controller-manager-76d5577b-qfp99\" (UID: \"914ce43c-e528-4921-b1b4-aa47cdf4b004\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.284798 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjt5x\" (UniqueName: \"kubernetes.io/projected/568c0ce9-12f3-4b04-9131-83771329a83e-kube-api-access-zjt5x\") pod \"placement-operator-controller-manager-66f6d6849b-mnzm7\" (UID: \"568c0ce9-12f3-4b04-9131-83771329a83e\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.290741 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-qfp99"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.315690 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmw7z\" (UniqueName: \"kubernetes.io/projected/413157c3-1aa7-425d-98ae-e44a7c19b425-kube-api-access-vmw7z\") pod \"octavia-operator-controller-manager-69f59f9d8-rtmss\" (UID: \"413157c3-1aa7-425d-98ae-e44a7c19b425\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.325262 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.326206 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.345006 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.364438 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.366712 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.370601 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.374388 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-hlhbr" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.381425 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.386045 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjt5x\" (UniqueName: \"kubernetes.io/projected/568c0ce9-12f3-4b04-9131-83771329a83e-kube-api-access-zjt5x\") pod \"placement-operator-controller-manager-66f6d6849b-mnzm7\" (UID: \"568c0ce9-12f3-4b04-9131-83771329a83e\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.386149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.386203 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74z48\" (UniqueName: \"kubernetes.io/projected/cb1fd16d-a191-4086-9a18-37b74183042d-kube-api-access-74z48\") pod \"ovn-operator-controller-manager-54d485fd9-29jxc\" (UID: \"cb1fd16d-a191-4086-9a18-37b74183042d\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.386225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpm6f\" (UniqueName: \"kubernetes.io/projected/f164aece-1b13-46b7-9da2-16967eb942f5-kube-api-access-tpm6f\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.386271 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrtc5\" (UniqueName: \"kubernetes.io/projected/914ce43c-e528-4921-b1b4-aa47cdf4b004-kube-api-access-hrtc5\") pod \"swift-operator-controller-manager-76d5577b-qfp99\" (UID: \"914ce43c-e528-4921-b1b4-aa47cdf4b004\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.386681 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.386770 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert podName:f164aece-1b13-46b7-9da2-16967eb942f5 nodeName:}" failed. No retries permitted until 2025-10-08 06:46:19.886748287 +0000 UTC m=+882.521188027 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert") pod "openstack-baremetal-operator-controller-manager-6875c66686lr5fk" (UID: "f164aece-1b13-46b7-9da2-16967eb942f5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.397693 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.399178 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.404020 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6n2vv" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.404184 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.408882 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjt5x\" (UniqueName: \"kubernetes.io/projected/568c0ce9-12f3-4b04-9131-83771329a83e-kube-api-access-zjt5x\") pod \"placement-operator-controller-manager-66f6d6849b-mnzm7\" (UID: \"568c0ce9-12f3-4b04-9131-83771329a83e\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.411263 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpm6f\" (UniqueName: \"kubernetes.io/projected/f164aece-1b13-46b7-9da2-16967eb942f5-kube-api-access-tpm6f\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.417857 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74z48\" (UniqueName: \"kubernetes.io/projected/cb1fd16d-a191-4086-9a18-37b74183042d-kube-api-access-74z48\") pod \"ovn-operator-controller-manager-54d485fd9-29jxc\" (UID: \"cb1fd16d-a191-4086-9a18-37b74183042d\") " pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.420186 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrtc5\" (UniqueName: \"kubernetes.io/projected/914ce43c-e528-4921-b1b4-aa47cdf4b004-kube-api-access-hrtc5\") pod \"swift-operator-controller-manager-76d5577b-qfp99\" (UID: \"914ce43c-e528-4921-b1b4-aa47cdf4b004\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.431493 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.431551 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.435036 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.436533 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.441201 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-hj69p" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.443671 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.447254 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.484418 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.489638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gdfg\" (UniqueName: \"kubernetes.io/projected/826e673f-b91d-4604-959f-098bcc2e6dcd-kube-api-access-4gdfg\") pod \"telemetry-operator-controller-manager-f589c7597-274pz\" (UID: \"826e673f-b91d-4604-959f-098bcc2e6dcd\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.489688 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm2f2\" (UniqueName: \"kubernetes.io/projected/17a2f9a4-8153-490e-be01-e452b10638e9-kube-api-access-fm2f2\") pod \"test-operator-controller-manager-6bb6dcddc-ldhbx\" (UID: \"17a2f9a4-8153-490e-be01-e452b10638e9\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.490908 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.491292 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.491366 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert podName:0b266b2d-a60b-4990-b720-800d655a4b5f nodeName:}" failed. No retries permitted until 2025-10-08 06:46:20.491325274 +0000 UTC m=+883.125765014 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert") pod "infra-operator-controller-manager-84788b6bc5-dqjfx" (UID: "0b266b2d-a60b-4990-b720-800d655a4b5f") : secret "infra-operator-webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.506382 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.508081 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.513272 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-4psbn" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.513565 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.519401 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.521503 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z"] Oct 08 06:46:19 crc kubenswrapper[4810]: W1008 06:46:19.526818 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c5ab097_4362_44a0_8f0a_a2a2483851f4.slice/crio-9d694e610401f84c9726b31e14b0c050379a1ac884a9d9c969014ac90826ab9f WatchSource:0}: Error finding container 9d694e610401f84c9726b31e14b0c050379a1ac884a9d9c969014ac90826ab9f: Status 404 returned error can't find the container with id 9d694e610401f84c9726b31e14b0c050379a1ac884a9d9c969014ac90826ab9f Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.550384 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.561115 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.562423 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.567030 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-n6crt" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.571338 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.593178 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gdfg\" (UniqueName: \"kubernetes.io/projected/826e673f-b91d-4604-959f-098bcc2e6dcd-kube-api-access-4gdfg\") pod \"telemetry-operator-controller-manager-f589c7597-274pz\" (UID: \"826e673f-b91d-4604-959f-098bcc2e6dcd\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.593253 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6cs2\" (UniqueName: \"kubernetes.io/projected/069894d9-bd50-4209-9df7-d941d44f0341-kube-api-access-h6cs2\") pod \"watcher-operator-controller-manager-5d98cc5575-ssc8r\" (UID: \"069894d9-bd50-4209-9df7-d941d44f0341\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.593279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm2f2\" (UniqueName: \"kubernetes.io/projected/17a2f9a4-8153-490e-be01-e452b10638e9-kube-api-access-fm2f2\") pod \"test-operator-controller-manager-6bb6dcddc-ldhbx\" (UID: \"17a2f9a4-8153-490e-be01-e452b10638e9\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.593338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea4ec203-527d-4deb-94c5-068053bbb7ee-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-ljndg\" (UID: \"ea4ec203-527d-4deb-94c5-068053bbb7ee\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.593394 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2djnc\" (UniqueName: \"kubernetes.io/projected/ea4ec203-527d-4deb-94c5-068053bbb7ee-kube-api-access-2djnc\") pod \"openstack-operator-controller-manager-6bfd56c677-ljndg\" (UID: \"ea4ec203-527d-4deb-94c5-068053bbb7ee\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.616400 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gdfg\" (UniqueName: \"kubernetes.io/projected/826e673f-b91d-4604-959f-098bcc2e6dcd-kube-api-access-4gdfg\") pod \"telemetry-operator-controller-manager-f589c7597-274pz\" (UID: \"826e673f-b91d-4604-959f-098bcc2e6dcd\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.627315 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm2f2\" (UniqueName: \"kubernetes.io/projected/17a2f9a4-8153-490e-be01-e452b10638e9-kube-api-access-fm2f2\") pod \"test-operator-controller-manager-6bb6dcddc-ldhbx\" (UID: \"17a2f9a4-8153-490e-be01-e452b10638e9\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.695188 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6cs2\" (UniqueName: \"kubernetes.io/projected/069894d9-bd50-4209-9df7-d941d44f0341-kube-api-access-h6cs2\") pod \"watcher-operator-controller-manager-5d98cc5575-ssc8r\" (UID: \"069894d9-bd50-4209-9df7-d941d44f0341\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.695270 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea4ec203-527d-4deb-94c5-068053bbb7ee-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-ljndg\" (UID: \"ea4ec203-527d-4deb-94c5-068053bbb7ee\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.695331 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2djnc\" (UniqueName: \"kubernetes.io/projected/ea4ec203-527d-4deb-94c5-068053bbb7ee-kube-api-access-2djnc\") pod \"openstack-operator-controller-manager-6bfd56c677-ljndg\" (UID: \"ea4ec203-527d-4deb-94c5-068053bbb7ee\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.695386 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8flsw\" (UniqueName: \"kubernetes.io/projected/133a94f9-3f74-42f7-95cf-dd0b193b6626-kube-api-access-8flsw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw\" (UID: \"133a94f9-3f74-42f7-95cf-dd0b193b6626\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.695993 4810 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.696065 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ea4ec203-527d-4deb-94c5-068053bbb7ee-cert podName:ea4ec203-527d-4deb-94c5-068053bbb7ee nodeName:}" failed. No retries permitted until 2025-10-08 06:46:20.196041535 +0000 UTC m=+882.830481275 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ea4ec203-527d-4deb-94c5-068053bbb7ee-cert") pod "openstack-operator-controller-manager-6bfd56c677-ljndg" (UID: "ea4ec203-527d-4deb-94c5-068053bbb7ee") : secret "webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.735711 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2djnc\" (UniqueName: \"kubernetes.io/projected/ea4ec203-527d-4deb-94c5-068053bbb7ee-kube-api-access-2djnc\") pod \"openstack-operator-controller-manager-6bfd56c677-ljndg\" (UID: \"ea4ec203-527d-4deb-94c5-068053bbb7ee\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.735914 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6cs2\" (UniqueName: \"kubernetes.io/projected/069894d9-bd50-4209-9df7-d941d44f0341-kube-api-access-h6cs2\") pod \"watcher-operator-controller-manager-5d98cc5575-ssc8r\" (UID: \"069894d9-bd50-4209-9df7-d941d44f0341\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.783615 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.796884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8flsw\" (UniqueName: \"kubernetes.io/projected/133a94f9-3f74-42f7-95cf-dd0b193b6626-kube-api-access-8flsw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw\" (UID: \"133a94f9-3f74-42f7-95cf-dd0b193b6626\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.797737 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.823210 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.823759 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8flsw\" (UniqueName: \"kubernetes.io/projected/133a94f9-3f74-42f7-95cf-dd0b193b6626-kube-api-access-8flsw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw\" (UID: \"133a94f9-3f74-42f7-95cf-dd0b193b6626\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.882393 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl"] Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.902498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.902675 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: E1008 06:46:19.902761 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert podName:f164aece-1b13-46b7-9da2-16967eb942f5 nodeName:}" failed. No retries permitted until 2025-10-08 06:46:20.902742462 +0000 UTC m=+883.537182202 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert") pod "openstack-baremetal-operator-controller-manager-6875c66686lr5fk" (UID: "f164aece-1b13-46b7-9da2-16967eb942f5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 06:46:19 crc kubenswrapper[4810]: I1008 06:46:19.917185 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.115459 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" event={"ID":"3fe805b1-a990-4683-8549-f73ebe9d850c","Type":"ContainerStarted","Data":"660a4cfc5fdbe6adfd43e0a432aa13c2f1008f01f7f68249e5082e4e1f8d67f1"} Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.116356 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" event={"ID":"2c5ab097-4362-44a0-8f0a-a2a2483851f4","Type":"ContainerStarted","Data":"9d694e610401f84c9726b31e14b0c050379a1ac884a9d9c969014ac90826ab9f"} Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.133023 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.156910 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.173020 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.193535 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.211369 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.212040 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea4ec203-527d-4deb-94c5-068053bbb7ee-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-ljndg\" (UID: \"ea4ec203-527d-4deb-94c5-068053bbb7ee\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.235408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea4ec203-527d-4deb-94c5-068053bbb7ee-cert\") pod \"openstack-operator-controller-manager-6bfd56c677-ljndg\" (UID: \"ea4ec203-527d-4deb-94c5-068053bbb7ee\") " pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:20 crc kubenswrapper[4810]: W1008 06:46:20.260983 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72a051e1_5079_45b0_b3d8_0592f872c85a.slice/crio-cd05fad7e1b4fcc33783b60898b7e3438ff120b492aa8d6e2ce5854a14b7194e WatchSource:0}: Error finding container cd05fad7e1b4fcc33783b60898b7e3438ff120b492aa8d6e2ce5854a14b7194e: Status 404 returned error can't find the container with id cd05fad7e1b4fcc33783b60898b7e3438ff120b492aa8d6e2ce5854a14b7194e Oct 08 06:46:20 crc kubenswrapper[4810]: W1008 06:46:20.337350 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24b64e0e_8eb4_4235_9086_8f1f61f2bd8e.slice/crio-0c44478bac60bebb8b6ed0803b46edd4361b0975a849d6a0b1d687902122c025 WatchSource:0}: Error finding container 0c44478bac60bebb8b6ed0803b46edd4361b0975a849d6a0b1d687902122c025: Status 404 returned error can't find the container with id 0c44478bac60bebb8b6ed0803b46edd4361b0975a849d6a0b1d687902122c025 Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.355937 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.386243 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.461326 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.465424 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.479245 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.491386 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.512010 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-qfp99"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.525179 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.532777 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.537086 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m"] Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.538292 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hrtc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-76d5577b-qfp99_openstack-operators(914ce43c-e528-4921-b1b4-aa47cdf4b004): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.540166 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b266b2d-a60b-4990-b720-800d655a4b5f-cert\") pod \"infra-operator-controller-manager-84788b6bc5-dqjfx\" (UID: \"0b266b2d-a60b-4990-b720-800d655a4b5f\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.546052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:20 crc kubenswrapper[4810]: W1008 06:46:20.550297 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod568c0ce9_12f3_4b04_9131_83771329a83e.slice/crio-6e49a6eeac329662fe367ecabf6992ae821a596407047b3f81c4d163dacc1b81 WatchSource:0}: Error finding container 6e49a6eeac329662fe367ecabf6992ae821a596407047b3f81c4d163dacc1b81: Status 404 returned error can't find the container with id 6e49a6eeac329662fe367ecabf6992ae821a596407047b3f81c4d163dacc1b81 Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.554788 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7"] Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.567520 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zjt5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-66f6d6849b-mnzm7_openstack-operators(568c0ce9-12f3-4b04-9131-83771329a83e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.690793 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw"] Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.699017 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fm2f2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-6bb6dcddc-ldhbx_openstack-operators(17a2f9a4-8153-490e-be01-e452b10638e9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.715511 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx"] Oct 08 06:46:20 crc kubenswrapper[4810]: W1008 06:46:20.723649 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod133a94f9_3f74_42f7_95cf_dd0b193b6626.slice/crio-b5a8845aeb101f2afcadcccd6b37261d2875119797d50a72476b6fd809c4d5d5 WatchSource:0}: Error finding container b5a8845aeb101f2afcadcccd6b37261d2875119797d50a72476b6fd809c4d5d5: Status 404 returned error can't find the container with id b5a8845aeb101f2afcadcccd6b37261d2875119797d50a72476b6fd809c4d5d5 Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.726406 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4gdfg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-f589c7597-274pz_openstack-operators(826e673f-b91d-4604-959f-098bcc2e6dcd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.727647 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8flsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw_openstack-operators(133a94f9-3f74-42f7-95cf-dd0b193b6626): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.730726 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" podUID="133a94f9-3f74-42f7-95cf-dd0b193b6626" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.732386 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.741850 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz"] Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.751487 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" podUID="914ce43c-e528-4921-b1b4-aa47cdf4b004" Oct 08 06:46:20 crc kubenswrapper[4810]: E1008 06:46:20.786866 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" podUID="568c0ce9-12f3-4b04-9131-83771329a83e" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.832896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg"] Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.943438 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:20 crc kubenswrapper[4810]: I1008 06:46:20.952490 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f164aece-1b13-46b7-9da2-16967eb942f5-cert\") pod \"openstack-baremetal-operator-controller-manager-6875c66686lr5fk\" (UID: \"f164aece-1b13-46b7-9da2-16967eb942f5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:21 crc kubenswrapper[4810]: E1008 06:46:21.048686 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" podUID="17a2f9a4-8153-490e-be01-e452b10638e9" Oct 08 06:46:21 crc kubenswrapper[4810]: E1008 06:46:21.054020 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" podUID="826e673f-b91d-4604-959f-098bcc2e6dcd" Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.104404 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx"] Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.129817 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" event={"ID":"133a94f9-3f74-42f7-95cf-dd0b193b6626","Type":"ContainerStarted","Data":"b5a8845aeb101f2afcadcccd6b37261d2875119797d50a72476b6fd809c4d5d5"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.136264 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" event={"ID":"413157c3-1aa7-425d-98ae-e44a7c19b425","Type":"ContainerStarted","Data":"f1d60f2083a1857d71e13b6105421c41dd4ffe7b0d1e45047fe1722f052cd878"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.140415 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" event={"ID":"ea4ec203-527d-4deb-94c5-068053bbb7ee","Type":"ContainerStarted","Data":"0347e15116334085e255c35b39d5b2a626cc41d1000bd4fb44536374d7207544"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.140445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" event={"ID":"ea4ec203-527d-4deb-94c5-068053bbb7ee","Type":"ContainerStarted","Data":"a927bcf875f5810a0948e887b8e5cc81dfc4f36272a6beb1080061186a290f1d"} Oct 08 06:46:21 crc kubenswrapper[4810]: E1008 06:46:21.141144 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" podUID="133a94f9-3f74-42f7-95cf-dd0b193b6626" Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.142693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" event={"ID":"e8deb064-ce04-4e12-ab87-e4853e499c20","Type":"ContainerStarted","Data":"dade920ff4d9ac2bee25e03d501d31abe5a8999bea4bb27eae2417d94d5d3462"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.145473 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" event={"ID":"17a2f9a4-8153-490e-be01-e452b10638e9","Type":"ContainerStarted","Data":"0656590d27bb5a8812773e7cee1a148f014584698f4b5b549fa6949cac0c1500"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.145510 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" event={"ID":"17a2f9a4-8153-490e-be01-e452b10638e9","Type":"ContainerStarted","Data":"7f05c4e3a66c8ec1d82677343f0036ab793188b88a750c71fe97d8e3106b3097"} Oct 08 06:46:21 crc kubenswrapper[4810]: E1008 06:46:21.147052 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" podUID="17a2f9a4-8153-490e-be01-e452b10638e9" Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.147651 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" event={"ID":"914ce43c-e528-4921-b1b4-aa47cdf4b004","Type":"ContainerStarted","Data":"968e2f0d5a4d76efd3ed9d5777d27ed981185c79b6949014c3fe62462e5f48ef"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.147679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" event={"ID":"914ce43c-e528-4921-b1b4-aa47cdf4b004","Type":"ContainerStarted","Data":"e1b89387eeba1f4ce62ccffd831e35d59f04969699422035cc8a4b0ffca22b57"} Oct 08 06:46:21 crc kubenswrapper[4810]: E1008 06:46:21.148754 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" podUID="914ce43c-e528-4921-b1b4-aa47cdf4b004" Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.173587 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" event={"ID":"72a051e1-5079-45b0-b3d8-0592f872c85a","Type":"ContainerStarted","Data":"cd05fad7e1b4fcc33783b60898b7e3438ff120b492aa8d6e2ce5854a14b7194e"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.208398 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" event={"ID":"c19839d0-64b9-4f39-a4e2-40cdc62bd9b1","Type":"ContainerStarted","Data":"4aaaba1caa8306c0c91086e9165b0ab869e55e3bbf5eb03f463ef59c0c411e91"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.215461 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" event={"ID":"24b64e0e-8eb4-4235-9086-8f1f61f2bd8e","Type":"ContainerStarted","Data":"0c44478bac60bebb8b6ed0803b46edd4361b0975a849d6a0b1d687902122c025"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.221610 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" event={"ID":"09c37aa7-fb5d-4984-9f32-86a4a323ea49","Type":"ContainerStarted","Data":"e82b8ad5944b52bfdcccf5f1a46eb4db315f3db3ff01d11501412f7e11b08cb1"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.223899 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" event={"ID":"75590e1d-5795-4ebe-afd0-9abb818efd63","Type":"ContainerStarted","Data":"fbcee3588871527c094d54596035f99a527b9e05ba99f82fb7219518274bf554"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.225437 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.229106 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" event={"ID":"568c0ce9-12f3-4b04-9131-83771329a83e","Type":"ContainerStarted","Data":"ab06dd6260e20752de30b9f870d2613f7346569122b5c525c2b25c7180e10078"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.229153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" event={"ID":"568c0ce9-12f3-4b04-9131-83771329a83e","Type":"ContainerStarted","Data":"6e49a6eeac329662fe367ecabf6992ae821a596407047b3f81c4d163dacc1b81"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.232024 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" event={"ID":"cb1fd16d-a191-4086-9a18-37b74183042d","Type":"ContainerStarted","Data":"7fd0b6f66b823142b3dfca627cb76ce4ab52bbdda06369aa6d293897055530f8"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.239926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" event={"ID":"681220bd-766a-4d1e-970b-662b4c64aaf6","Type":"ContainerStarted","Data":"bc916bc9e7859e5a72914202a61591dd7675f3daf601ca6d9d27abb34647ca72"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.246142 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" event={"ID":"6fc03514-36ec-4d41-8ad1-306ff2767a2f","Type":"ContainerStarted","Data":"a2455e804c55922517647e7278dc3d90898e2b2b449b5d6fec8969d40ec83a9c"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.251579 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" event={"ID":"069894d9-bd50-4209-9df7-d941d44f0341","Type":"ContainerStarted","Data":"af88b1e71807866c612b6e9b31e0e4e3ef0152cdfb143998c9e4dc533d8f1e34"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.260632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" event={"ID":"826e673f-b91d-4604-959f-098bcc2e6dcd","Type":"ContainerStarted","Data":"1da438ede255412810aca5a48250bc7bf6922dda01878df6c9beb87f228e8a12"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.260683 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" event={"ID":"826e673f-b91d-4604-959f-098bcc2e6dcd","Type":"ContainerStarted","Data":"463bd69b48916e32fa74099a7f625dbd1636b06c0eb24971f627d15adefc4db5"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.263227 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" event={"ID":"8181e145-9c88-47f0-a9a8-dc03947b75ef","Type":"ContainerStarted","Data":"cc2fec5f7f59591761ab8465cdcc95e71fb2651c44dba388bfed027aedf93e92"} Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.273031 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" event={"ID":"0bebe3c8-1035-40b4-b533-4c214423f6b2","Type":"ContainerStarted","Data":"1ceaaad6f1080c3d72a85f0e1a9f207f032fc14832a589083534313215d33ed6"} Oct 08 06:46:21 crc kubenswrapper[4810]: E1008 06:46:21.289690 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" podUID="568c0ce9-12f3-4b04-9131-83771329a83e" Oct 08 06:46:21 crc kubenswrapper[4810]: E1008 06:46:21.289721 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" podUID="826e673f-b91d-4604-959f-098bcc2e6dcd" Oct 08 06:46:21 crc kubenswrapper[4810]: I1008 06:46:21.958314 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk"] Oct 08 06:46:21 crc kubenswrapper[4810]: W1008 06:46:21.980199 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf164aece_1b13_46b7_9da2_16967eb942f5.slice/crio-839ba568003a80b500398870c9bdc130aba6379d5d6e64ea15b18cdc165bd215 WatchSource:0}: Error finding container 839ba568003a80b500398870c9bdc130aba6379d5d6e64ea15b18cdc165bd215: Status 404 returned error can't find the container with id 839ba568003a80b500398870c9bdc130aba6379d5d6e64ea15b18cdc165bd215 Oct 08 06:46:22 crc kubenswrapper[4810]: I1008 06:46:22.288127 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" event={"ID":"0b266b2d-a60b-4990-b720-800d655a4b5f","Type":"ContainerStarted","Data":"7fe6ca43f8d6392669089d5656e67a6322df855f87768d0ca70b55e1e39be52e"} Oct 08 06:46:22 crc kubenswrapper[4810]: I1008 06:46:22.293333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" event={"ID":"f164aece-1b13-46b7-9da2-16967eb942f5","Type":"ContainerStarted","Data":"839ba568003a80b500398870c9bdc130aba6379d5d6e64ea15b18cdc165bd215"} Oct 08 06:46:22 crc kubenswrapper[4810]: I1008 06:46:22.308591 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" event={"ID":"ea4ec203-527d-4deb-94c5-068053bbb7ee","Type":"ContainerStarted","Data":"42233822c7e2c8de6b43d9ab13169478f31a68be33f64a2dfae721dd2c3b8797"} Oct 08 06:46:22 crc kubenswrapper[4810]: I1008 06:46:22.308754 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:22 crc kubenswrapper[4810]: E1008 06:46:22.310715 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" podUID="17a2f9a4-8153-490e-be01-e452b10638e9" Oct 08 06:46:22 crc kubenswrapper[4810]: E1008 06:46:22.312062 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" podUID="568c0ce9-12f3-4b04-9131-83771329a83e" Oct 08 06:46:22 crc kubenswrapper[4810]: E1008 06:46:22.312592 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" podUID="133a94f9-3f74-42f7-95cf-dd0b193b6626" Oct 08 06:46:22 crc kubenswrapper[4810]: E1008 06:46:22.314940 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" podUID="826e673f-b91d-4604-959f-098bcc2e6dcd" Oct 08 06:46:22 crc kubenswrapper[4810]: E1008 06:46:22.333316 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" podUID="914ce43c-e528-4921-b1b4-aa47cdf4b004" Oct 08 06:46:22 crc kubenswrapper[4810]: I1008 06:46:22.372601 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" podStartSLOduration=3.372585283 podStartE2EDuration="3.372585283s" podCreationTimestamp="2025-10-08 06:46:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:46:22.36372087 +0000 UTC m=+884.998160610" watchObservedRunningTime="2025-10-08 06:46:22.372585283 +0000 UTC m=+885.007025023" Oct 08 06:46:30 crc kubenswrapper[4810]: I1008 06:46:30.473573 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6bfd56c677-ljndg" Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.453394 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" event={"ID":"069894d9-bd50-4209-9df7-d941d44f0341","Type":"ContainerStarted","Data":"83ac00b0be6302924e20a0a7866049237ff1b065993e2afa4845ae2d879a528b"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.487167 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" event={"ID":"3fe805b1-a990-4683-8549-f73ebe9d850c","Type":"ContainerStarted","Data":"c754ea3d76c7a8408a1746cd3e01edf931938693ebe0a7a77c7569f528f3466d"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.522741 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" event={"ID":"75590e1d-5795-4ebe-afd0-9abb818efd63","Type":"ContainerStarted","Data":"1cc547b9db91a81a880a729098e7efdeb1db0e325dffee552ef69e6a29181316"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.539545 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" event={"ID":"0bebe3c8-1035-40b4-b533-4c214423f6b2","Type":"ContainerStarted","Data":"e602222b3b2d124c0434ed2b4ddb748626f36fbfea620ec3b4952a22ea4eb7ff"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.573544 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" event={"ID":"cb1fd16d-a191-4086-9a18-37b74183042d","Type":"ContainerStarted","Data":"66622d77aed4c6ae8b1d58f34852fa7ddfbcb8857d30b9ed933f91f763ca150f"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.595275 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" event={"ID":"72a051e1-5079-45b0-b3d8-0592f872c85a","Type":"ContainerStarted","Data":"83c64732df053c3569138bce4f2c28020dceacc230607aade5c48de12fd82acf"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.621219 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" event={"ID":"c19839d0-64b9-4f39-a4e2-40cdc62bd9b1","Type":"ContainerStarted","Data":"d34e4421aa65d656a72249bc5193bcd634379525a25d171367fca8b150aca227"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.642279 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" event={"ID":"8181e145-9c88-47f0-a9a8-dc03947b75ef","Type":"ContainerStarted","Data":"dea04a31598fe398399c7d2dac2fac51faa361ab1f47e99a29d54d52e08f6b94"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.642623 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.665039 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" event={"ID":"e8deb064-ce04-4e12-ab87-e4853e499c20","Type":"ContainerStarted","Data":"b4c5644d833d10caceb56838a8b76b681be18082bcfb1b3abf3b53c11d21461f"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.681829 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" event={"ID":"24b64e0e-8eb4-4235-9086-8f1f61f2bd8e","Type":"ContainerStarted","Data":"f33f326cac06e519f2b0234401c0ea6789a12d3e8dd65b2a8c4e92583d1540c3"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.693271 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" podStartSLOduration=3.318323275 podStartE2EDuration="15.69325243s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.222202279 +0000 UTC m=+882.856642019" lastFinishedPulling="2025-10-08 06:46:32.597131434 +0000 UTC m=+895.231571174" observedRunningTime="2025-10-08 06:46:33.691539424 +0000 UTC m=+896.325979174" watchObservedRunningTime="2025-10-08 06:46:33.69325243 +0000 UTC m=+896.327692170" Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.709298 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" event={"ID":"09c37aa7-fb5d-4984-9f32-86a4a323ea49","Type":"ContainerStarted","Data":"7f750b1df70d7e5d58ebcba9c866ea6c05c88abf60dc20f90c997828a4eb582b"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.735618 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" event={"ID":"0b266b2d-a60b-4990-b720-800d655a4b5f","Type":"ContainerStarted","Data":"e9c64e59fa18b157831585245643155f601363c7e55b0387c180ffc5e57279d7"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.742172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" event={"ID":"f164aece-1b13-46b7-9da2-16967eb942f5","Type":"ContainerStarted","Data":"fe55b8f8d6f56cac096b8da7e22b0f2cd07a15c190c6ae8fd7a50758988d5e6f"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.761932 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" event={"ID":"6fc03514-36ec-4d41-8ad1-306ff2767a2f","Type":"ContainerStarted","Data":"54aed6f3edeceaf8cc8c5c85d2b39a1cc003b4c1c9f56e57fff4d29932d50b8f"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.776141 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" event={"ID":"2c5ab097-4362-44a0-8f0a-a2a2483851f4","Type":"ContainerStarted","Data":"b44939270e9cb2b61b5dfa2d06e579f4a8773174375922e4378a6e625aae11d2"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.786464 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" event={"ID":"413157c3-1aa7-425d-98ae-e44a7c19b425","Type":"ContainerStarted","Data":"78c001db4cfa024a7709225813d4af4012d2e0480075c45f3a871fb7c1dc5706"} Oct 08 06:46:33 crc kubenswrapper[4810]: I1008 06:46:33.824590 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" event={"ID":"681220bd-766a-4d1e-970b-662b4c64aaf6","Type":"ContainerStarted","Data":"0703f33fc5b7486c8f40a5f536383c4f2ec0387a1330715c9ccb6e583eb5deef"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.835816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" event={"ID":"cb1fd16d-a191-4086-9a18-37b74183042d","Type":"ContainerStarted","Data":"830cbcdc7e1b80967d2bc80d0e91c9b5a555dae8ac58dd8254c830435e0743c4"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.836383 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.840867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" event={"ID":"0b266b2d-a60b-4990-b720-800d655a4b5f","Type":"ContainerStarted","Data":"1a7aec9edfcfbafd24269768c159c5b373f2fadfb3b49afcc51fabd1a07e5a1c"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.840993 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.859884 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" event={"ID":"f164aece-1b13-46b7-9da2-16967eb942f5","Type":"ContainerStarted","Data":"ab0dd8b13091adef714c0a89feb67b0870c061b87af5c3ff3bb7526b6aa62571"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.860936 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.877480 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" event={"ID":"6fc03514-36ec-4d41-8ad1-306ff2767a2f","Type":"ContainerStarted","Data":"9de9fb72f471799f70655f369104bdab9c11a17e68c3d0e681cb9321d956d76d"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.877698 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.880512 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" event={"ID":"c19839d0-64b9-4f39-a4e2-40cdc62bd9b1","Type":"ContainerStarted","Data":"79f5b25028163a4028d41d86d372a4c40ed2ad549f6d558649e5aea7364a3ffe"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.881743 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.895035 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" event={"ID":"3fe805b1-a990-4683-8549-f73ebe9d850c","Type":"ContainerStarted","Data":"556de4b457d144a30d8fed6282fb9fa3ffd1c14bbff2d3a50e11eef4c6d2b8e4"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.898796 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" podStartSLOduration=4.716627444 podStartE2EDuration="16.898755655s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.48346289 +0000 UTC m=+883.117902640" lastFinishedPulling="2025-10-08 06:46:32.665591111 +0000 UTC m=+895.300030851" observedRunningTime="2025-10-08 06:46:34.854846451 +0000 UTC m=+897.489286191" watchObservedRunningTime="2025-10-08 06:46:34.898755655 +0000 UTC m=+897.533195395" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.901426 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.906833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" event={"ID":"8181e145-9c88-47f0-a9a8-dc03947b75ef","Type":"ContainerStarted","Data":"e79e0d965d67784b9cd9f0a487a26ceb5c4366dd50734c832f62a78b1e501534"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.905831 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" podStartSLOduration=5.485225643 podStartE2EDuration="16.905802018s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:21.124531813 +0000 UTC m=+883.758971553" lastFinishedPulling="2025-10-08 06:46:32.545108188 +0000 UTC m=+895.179547928" observedRunningTime="2025-10-08 06:46:34.884612877 +0000 UTC m=+897.519052607" watchObservedRunningTime="2025-10-08 06:46:34.905802018 +0000 UTC m=+897.540241758" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.928263 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" podStartSLOduration=4.584443031 podStartE2EDuration="16.928241193s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.170130171 +0000 UTC m=+882.804569911" lastFinishedPulling="2025-10-08 06:46:32.513928333 +0000 UTC m=+895.148368073" observedRunningTime="2025-10-08 06:46:34.908465701 +0000 UTC m=+897.542905441" watchObservedRunningTime="2025-10-08 06:46:34.928241193 +0000 UTC m=+897.562680933" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.937026 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" event={"ID":"0bebe3c8-1035-40b4-b533-4c214423f6b2","Type":"ContainerStarted","Data":"0a4815cb933eb88fd4559538af735bd91b7d8f4fc8f62c4e1acc7551b5c45215"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.938088 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.946589 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" podStartSLOduration=4.868838166 podStartE2EDuration="16.946559255s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.537046568 +0000 UTC m=+883.171486308" lastFinishedPulling="2025-10-08 06:46:32.614767627 +0000 UTC m=+895.249207397" observedRunningTime="2025-10-08 06:46:34.938782172 +0000 UTC m=+897.573221912" watchObservedRunningTime="2025-10-08 06:46:34.946559255 +0000 UTC m=+897.580998996" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.951612 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" event={"ID":"069894d9-bd50-4209-9df7-d941d44f0341","Type":"ContainerStarted","Data":"e46cfc0fde46f664f0e104630ae628373a81a39b9e11b7e7302dd8f7c41b28a1"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.951865 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.968216 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" event={"ID":"09c37aa7-fb5d-4984-9f32-86a4a323ea49","Type":"ContainerStarted","Data":"9babd943f28116bc7df9d27a30041cebc2fccb899fd5b18a197c47262bc2d44e"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.969387 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.986672 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" event={"ID":"681220bd-766a-4d1e-970b-662b4c64aaf6","Type":"ContainerStarted","Data":"119a00454e0c01d81edcbf733ae711f4dfc189a57445f35770fa8e957c472fbb"} Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.987026 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" Oct 08 06:46:34 crc kubenswrapper[4810]: I1008 06:46:34.997837 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" podStartSLOduration=6.305051765 podStartE2EDuration="16.99781985s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:21.98465656 +0000 UTC m=+884.619096300" lastFinishedPulling="2025-10-08 06:46:32.677424645 +0000 UTC m=+895.311864385" observedRunningTime="2025-10-08 06:46:34.995415174 +0000 UTC m=+897.629854914" watchObservedRunningTime="2025-10-08 06:46:34.99781985 +0000 UTC m=+897.632259590" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.002599 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" event={"ID":"75590e1d-5795-4ebe-afd0-9abb818efd63","Type":"ContainerStarted","Data":"9a895673136ef14666dd5f87b4c50eaa5b360f3dbe93cf9223ac3816e90710a2"} Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.003818 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.037497 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" event={"ID":"72a051e1-5079-45b0-b3d8-0592f872c85a","Type":"ContainerStarted","Data":"c5ae0ad561ba9dfe3ebdd705aade15922da6940eeacec458223dd4f6f35dccf4"} Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.038679 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.043579 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" podStartSLOduration=4.737790865 podStartE2EDuration="17.043556304s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.283428807 +0000 UTC m=+882.917868547" lastFinishedPulling="2025-10-08 06:46:32.589194226 +0000 UTC m=+895.223633986" observedRunningTime="2025-10-08 06:46:35.038196148 +0000 UTC m=+897.672635888" watchObservedRunningTime="2025-10-08 06:46:35.043556304 +0000 UTC m=+897.677996044" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.049616 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" event={"ID":"2c5ab097-4362-44a0-8f0a-a2a2483851f4","Type":"ContainerStarted","Data":"e3a07fc3facdfc0915ec55c6e35ee70cd2760da77ec293e24903c6efc73ce495"} Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.050769 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.103299 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" event={"ID":"e8deb064-ce04-4e12-ab87-e4853e499c20","Type":"ContainerStarted","Data":"d3abd06c760a14af0622cb150f2f61abf3e9e6fb3192ef414b58b3a4fb3b51ed"} Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.104259 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.116330 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" podStartSLOduration=4.989126614 podStartE2EDuration="17.116300528s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.537134301 +0000 UTC m=+883.171574041" lastFinishedPulling="2025-10-08 06:46:32.664308215 +0000 UTC m=+895.298747955" observedRunningTime="2025-10-08 06:46:35.115509506 +0000 UTC m=+897.749949256" watchObservedRunningTime="2025-10-08 06:46:35.116300528 +0000 UTC m=+897.750740268" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.123246 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" podStartSLOduration=4.389443686 podStartE2EDuration="17.123223298s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:19.919061889 +0000 UTC m=+882.553501629" lastFinishedPulling="2025-10-08 06:46:32.652841491 +0000 UTC m=+895.287281241" observedRunningTime="2025-10-08 06:46:35.072194919 +0000 UTC m=+897.706634669" watchObservedRunningTime="2025-10-08 06:46:35.123223298 +0000 UTC m=+897.757663028" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.149318 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" event={"ID":"24b64e0e-8eb4-4235-9086-8f1f61f2bd8e","Type":"ContainerStarted","Data":"75de2710977b32a210a979e35d1141d87c7569706d0795b7753c6cf959b4d2ec"} Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.150480 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.159223 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" event={"ID":"413157c3-1aa7-425d-98ae-e44a7c19b425","Type":"ContainerStarted","Data":"cc6fa7c979aeb8bb507effb7dec92d45fbb36db25f4d785a5f39e7ec1e443d25"} Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.159782 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.173192 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" podStartSLOduration=5.149836299 podStartE2EDuration="17.173169187s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.575994566 +0000 UTC m=+883.210434306" lastFinishedPulling="2025-10-08 06:46:32.599327454 +0000 UTC m=+895.233767194" observedRunningTime="2025-10-08 06:46:35.170378761 +0000 UTC m=+897.804818501" watchObservedRunningTime="2025-10-08 06:46:35.173169187 +0000 UTC m=+897.807608927" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.208180 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" podStartSLOduration=4.286088944 podStartE2EDuration="16.208151376s" podCreationTimestamp="2025-10-08 06:46:19 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.783832463 +0000 UTC m=+883.418272203" lastFinishedPulling="2025-10-08 06:46:32.705894895 +0000 UTC m=+895.340334635" observedRunningTime="2025-10-08 06:46:35.206580593 +0000 UTC m=+897.841020333" watchObservedRunningTime="2025-10-08 06:46:35.208151376 +0000 UTC m=+897.842591116" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.240663 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" podStartSLOduration=5.046566089 podStartE2EDuration="17.240639757s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.483355247 +0000 UTC m=+883.117794987" lastFinishedPulling="2025-10-08 06:46:32.677428905 +0000 UTC m=+895.311868655" observedRunningTime="2025-10-08 06:46:35.237443919 +0000 UTC m=+897.871883659" watchObservedRunningTime="2025-10-08 06:46:35.240639757 +0000 UTC m=+897.875079497" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.334010 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" podStartSLOduration=4.387820241 podStartE2EDuration="17.333995435s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:19.60354563 +0000 UTC m=+882.237985370" lastFinishedPulling="2025-10-08 06:46:32.549720824 +0000 UTC m=+895.184160564" observedRunningTime="2025-10-08 06:46:35.296170509 +0000 UTC m=+897.930610279" watchObservedRunningTime="2025-10-08 06:46:35.333995435 +0000 UTC m=+897.968435175" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.339283 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" podStartSLOduration=5.162212908 podStartE2EDuration="17.33927002s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.352350376 +0000 UTC m=+882.986790116" lastFinishedPulling="2025-10-08 06:46:32.529407488 +0000 UTC m=+895.163847228" observedRunningTime="2025-10-08 06:46:35.337538252 +0000 UTC m=+897.971977992" watchObservedRunningTime="2025-10-08 06:46:35.33927002 +0000 UTC m=+897.973709760" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.363489 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" podStartSLOduration=4.95029688 podStartE2EDuration="17.363471194s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.25291031 +0000 UTC m=+882.887350050" lastFinishedPulling="2025-10-08 06:46:32.666084624 +0000 UTC m=+895.300524364" observedRunningTime="2025-10-08 06:46:35.361652154 +0000 UTC m=+897.996091904" watchObservedRunningTime="2025-10-08 06:46:35.363471194 +0000 UTC m=+897.997910934" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.389024 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" podStartSLOduration=5.154096407 podStartE2EDuration="17.389005824s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.365779794 +0000 UTC m=+883.000219534" lastFinishedPulling="2025-10-08 06:46:32.600689211 +0000 UTC m=+895.235128951" observedRunningTime="2025-10-08 06:46:35.385686903 +0000 UTC m=+898.020126643" watchObservedRunningTime="2025-10-08 06:46:35.389005824 +0000 UTC m=+898.023445564" Oct 08 06:46:35 crc kubenswrapper[4810]: I1008 06:46:35.419568 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" podStartSLOduration=5.087615763 podStartE2EDuration="17.41952775s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.282957443 +0000 UTC m=+882.917397183" lastFinishedPulling="2025-10-08 06:46:32.61486943 +0000 UTC m=+895.249309170" observedRunningTime="2025-10-08 06:46:35.406857723 +0000 UTC m=+898.041297473" watchObservedRunningTime="2025-10-08 06:46:35.41952775 +0000 UTC m=+898.053967490" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.197671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" event={"ID":"568c0ce9-12f3-4b04-9131-83771329a83e","Type":"ContainerStarted","Data":"e79a6a50778d49418b0f013c43fc2cfaac684d9165a233e3dd5acc54a4dffbde"} Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.198410 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.201217 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" event={"ID":"914ce43c-e528-4921-b1b4-aa47cdf4b004","Type":"ContainerStarted","Data":"73d0fd702b1eda74d4e0fafe61b622d6147cde9c81c58cbffed4f7240c5407d4"} Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.201557 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.210043 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" event={"ID":"17a2f9a4-8153-490e-be01-e452b10638e9","Type":"ContainerStarted","Data":"cffb0652040bf37d9f3b6e7d06199a5fdcb93591e6f724acf21236378156e109"} Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.212931 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.228239 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" podStartSLOduration=2.938976795 podStartE2EDuration="20.22821375s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.567301018 +0000 UTC m=+883.201740758" lastFinishedPulling="2025-10-08 06:46:37.856537963 +0000 UTC m=+900.490977713" observedRunningTime="2025-10-08 06:46:38.220566551 +0000 UTC m=+900.855006301" watchObservedRunningTime="2025-10-08 06:46:38.22821375 +0000 UTC m=+900.862653490" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.246433 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" podStartSLOduration=3.087994941 podStartE2EDuration="20.246408259s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.698850854 +0000 UTC m=+883.333290594" lastFinishedPulling="2025-10-08 06:46:37.857264172 +0000 UTC m=+900.491703912" observedRunningTime="2025-10-08 06:46:38.23730139 +0000 UTC m=+900.871741140" watchObservedRunningTime="2025-10-08 06:46:38.246408259 +0000 UTC m=+900.880847999" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.263415 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" podStartSLOduration=2.942998237 podStartE2EDuration="20.263385975s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.537279915 +0000 UTC m=+883.171719655" lastFinishedPulling="2025-10-08 06:46:37.857667653 +0000 UTC m=+900.492107393" observedRunningTime="2025-10-08 06:46:38.256622839 +0000 UTC m=+900.891062579" watchObservedRunningTime="2025-10-08 06:46:38.263385975 +0000 UTC m=+900.897825715" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.837587 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-2lq4z" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.855224 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f56ff694-tlndl" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.863878 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-h4wk6" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.889533 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-fd648f65-nwwg7" Oct 08 06:46:38 crc kubenswrapper[4810]: I1008 06:46:38.917791 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-7ccfc8cf49-ntp7k" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.017280 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b477879bc-8jhkm" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.120804 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5467f8988c-s4m9g" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.172248 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b84cc7657-9gdl9" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.235255 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" event={"ID":"133a94f9-3f74-42f7-95cf-dd0b193b6626","Type":"ContainerStarted","Data":"a741d72d8a9404909c4bf4c80ed5334acab8ba4ec9a143699136197f55d4b2e4"} Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.255457 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw" podStartSLOduration=3.134593819 podStartE2EDuration="20.255436408s" podCreationTimestamp="2025-10-08 06:46:19 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.727467988 +0000 UTC m=+883.361907728" lastFinishedPulling="2025-10-08 06:46:37.848310547 +0000 UTC m=+900.482750317" observedRunningTime="2025-10-08 06:46:39.25293574 +0000 UTC m=+901.887375480" watchObservedRunningTime="2025-10-08 06:46:39.255436408 +0000 UTC m=+901.889876148" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.268529 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-657k2" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.294090 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-ckvcg" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.331475 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-8cmwf" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.332604 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-r9g7m" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.385430 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-rtmss" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.455916 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54d485fd9-29jxc" Oct 08 06:46:39 crc kubenswrapper[4810]: I1008 06:46:39.828434 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-ssc8r" Oct 08 06:46:40 crc kubenswrapper[4810]: I1008 06:46:40.249176 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" event={"ID":"826e673f-b91d-4604-959f-098bcc2e6dcd","Type":"ContainerStarted","Data":"6b2e1b68aaa209ec4e797d2e6ae18f0e754a404ec60ea440d6e41f900a87d7e1"} Oct 08 06:46:40 crc kubenswrapper[4810]: I1008 06:46:40.249535 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" Oct 08 06:46:40 crc kubenswrapper[4810]: I1008 06:46:40.273227 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" podStartSLOduration=3.199650591 podStartE2EDuration="22.273202736s" podCreationTimestamp="2025-10-08 06:46:18 +0000 UTC" firstStartedPulling="2025-10-08 06:46:20.726209584 +0000 UTC m=+883.360649364" lastFinishedPulling="2025-10-08 06:46:39.799761769 +0000 UTC m=+902.434201509" observedRunningTime="2025-10-08 06:46:40.270638306 +0000 UTC m=+902.905078066" watchObservedRunningTime="2025-10-08 06:46:40.273202736 +0000 UTC m=+902.907642476" Oct 08 06:46:40 crc kubenswrapper[4810]: I1008 06:46:40.558350 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-dqjfx" Oct 08 06:46:41 crc kubenswrapper[4810]: I1008 06:46:41.236404 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6875c66686lr5fk" Oct 08 06:46:49 crc kubenswrapper[4810]: I1008 06:46:49.432112 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:46:49 crc kubenswrapper[4810]: I1008 06:46:49.432633 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:46:49 crc kubenswrapper[4810]: I1008 06:46:49.489329 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-mnzm7" Oct 08 06:46:49 crc kubenswrapper[4810]: I1008 06:46:49.524294 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-76d5577b-qfp99" Oct 08 06:46:49 crc kubenswrapper[4810]: I1008 06:46:49.787460 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-274pz" Oct 08 06:46:49 crc kubenswrapper[4810]: I1008 06:46:49.802206 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-ldhbx" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.812321 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-2j9fh"] Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.823865 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.829699 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.829882 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-nhrtr" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.832459 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.832797 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.843553 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-2j9fh"] Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.888588 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adaacb32-8c3c-4eb7-a376-a60927949bb9-config\") pod \"dnsmasq-dns-7bfcb9d745-2j9fh\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.888858 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w2d6\" (UniqueName: \"kubernetes.io/projected/adaacb32-8c3c-4eb7-a376-a60927949bb9-kube-api-access-6w2d6\") pod \"dnsmasq-dns-7bfcb9d745-2j9fh\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.898572 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-p4g8n"] Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.899733 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.904013 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.924808 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-p4g8n"] Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.991185 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w2d6\" (UniqueName: \"kubernetes.io/projected/adaacb32-8c3c-4eb7-a376-a60927949bb9-kube-api-access-6w2d6\") pod \"dnsmasq-dns-7bfcb9d745-2j9fh\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.991290 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-config\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.991334 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq5jg\" (UniqueName: \"kubernetes.io/projected/54cef57a-8119-4756-bea5-09f819e851b9-kube-api-access-xq5jg\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.991367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adaacb32-8c3c-4eb7-a376-a60927949bb9-config\") pod \"dnsmasq-dns-7bfcb9d745-2j9fh\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.991410 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-dns-svc\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:04 crc kubenswrapper[4810]: I1008 06:47:04.992881 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adaacb32-8c3c-4eb7-a376-a60927949bb9-config\") pod \"dnsmasq-dns-7bfcb9d745-2j9fh\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.026929 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w2d6\" (UniqueName: \"kubernetes.io/projected/adaacb32-8c3c-4eb7-a376-a60927949bb9-kube-api-access-6w2d6\") pod \"dnsmasq-dns-7bfcb9d745-2j9fh\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.093123 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-config\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.093203 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq5jg\" (UniqueName: \"kubernetes.io/projected/54cef57a-8119-4756-bea5-09f819e851b9-kube-api-access-xq5jg\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.093256 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-dns-svc\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.094634 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-dns-svc\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.094945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-config\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.114601 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq5jg\" (UniqueName: \"kubernetes.io/projected/54cef57a-8119-4756-bea5-09f819e851b9-kube-api-access-xq5jg\") pod \"dnsmasq-dns-758b79db4c-p4g8n\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.153615 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.231855 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.687341 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-2j9fh"] Oct 08 06:47:05 crc kubenswrapper[4810]: I1008 06:47:05.783191 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-p4g8n"] Oct 08 06:47:05 crc kubenswrapper[4810]: W1008 06:47:05.790710 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cef57a_8119_4756_bea5_09f819e851b9.slice/crio-071179ba28a7432d6bf5571c8d9940c38024601fb12f5024da982c4e43dd0d65 WatchSource:0}: Error finding container 071179ba28a7432d6bf5571c8d9940c38024601fb12f5024da982c4e43dd0d65: Status 404 returned error can't find the container with id 071179ba28a7432d6bf5571c8d9940c38024601fb12f5024da982c4e43dd0d65 Oct 08 06:47:06 crc kubenswrapper[4810]: I1008 06:47:06.523556 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" event={"ID":"adaacb32-8c3c-4eb7-a376-a60927949bb9","Type":"ContainerStarted","Data":"0f0fe5d95e6c526565b0d2f7a06adb42781ba0b2091b7da5919648e8502934ac"} Oct 08 06:47:06 crc kubenswrapper[4810]: I1008 06:47:06.528826 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" event={"ID":"54cef57a-8119-4756-bea5-09f819e851b9","Type":"ContainerStarted","Data":"071179ba28a7432d6bf5571c8d9940c38024601fb12f5024da982c4e43dd0d65"} Oct 08 06:47:06 crc kubenswrapper[4810]: I1008 06:47:06.962556 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-p4g8n"] Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.038710 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-644597f84c-krjcl"] Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.047375 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.056671 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-krjcl"] Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.153455 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvsrd\" (UniqueName: \"kubernetes.io/projected/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-kube-api-access-rvsrd\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.153547 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-dns-svc\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.153585 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-config\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.254900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-config\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.255748 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvsrd\" (UniqueName: \"kubernetes.io/projected/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-kube-api-access-rvsrd\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.255928 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-dns-svc\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.257686 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-config\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.261799 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-dns-svc\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.287120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvsrd\" (UniqueName: \"kubernetes.io/projected/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-kube-api-access-rvsrd\") pod \"dnsmasq-dns-644597f84c-krjcl\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.479534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.596262 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-2j9fh"] Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.625949 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-cq9qj"] Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.631759 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.653871 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-cq9qj"] Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.767687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjtgg\" (UniqueName: \"kubernetes.io/projected/e046bc15-ade9-481c-a5c0-9c01f4794f36-kube-api-access-xjtgg\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.772238 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-config\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.772399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-dns-svc\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.874135 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-config\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.874263 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-dns-svc\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.874332 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjtgg\" (UniqueName: \"kubernetes.io/projected/e046bc15-ade9-481c-a5c0-9c01f4794f36-kube-api-access-xjtgg\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.875234 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-config\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.875563 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-dns-svc\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:07 crc kubenswrapper[4810]: I1008 06:47:07.909486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjtgg\" (UniqueName: \"kubernetes.io/projected/e046bc15-ade9-481c-a5c0-9c01f4794f36-kube-api-access-xjtgg\") pod \"dnsmasq-dns-77597f887-cq9qj\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.040575 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.171734 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-krjcl"] Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.215851 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.218375 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.220927 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.221320 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.221453 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ns45r" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.221651 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.221848 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.222182 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.222443 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.244095 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.351549 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-cq9qj"] Oct 08 06:47:08 crc kubenswrapper[4810]: W1008 06:47:08.372383 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode046bc15_ade9_481c_a5c0_9c01f4794f36.slice/crio-98191b8b4060a38247915b86bc963821f84618c63106ba35a6b155009ab40f43 WatchSource:0}: Error finding container 98191b8b4060a38247915b86bc963821f84618c63106ba35a6b155009ab40f43: Status 404 returned error can't find the container with id 98191b8b4060a38247915b86bc963821f84618c63106ba35a6b155009ab40f43 Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395118 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395176 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395398 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395434 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395470 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395519 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395556 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm62m\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-kube-api-access-xm62m\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395703 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.395747 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497682 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497775 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497811 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497843 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497873 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm62m\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-kube-api-access-xm62m\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497910 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497958 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.497994 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.498022 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.498044 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.500686 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.502118 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.502504 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.503236 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.503819 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.506747 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.509729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.514014 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.514016 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.514105 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.522694 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm62m\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-kube-api-access-xm62m\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.543246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.565334 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.583145 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-krjcl" event={"ID":"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef","Type":"ContainerStarted","Data":"65d1515a7ea2337a45c74c7a240ee260cf60f3c6b3077371578a2b385addb848"} Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.586649 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-cq9qj" event={"ID":"e046bc15-ade9-481c-a5c0-9c01f4794f36","Type":"ContainerStarted","Data":"98191b8b4060a38247915b86bc963821f84618c63106ba35a6b155009ab40f43"} Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.788171 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.792998 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.799498 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.800180 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.800319 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ffw96" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.800591 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.800646 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.800763 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.800901 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.804577 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.907228 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908009 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908035 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4deb0612-547f-4067-b95b-5794663d21aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908094 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhz55\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-kube-api-access-zhz55\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908419 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908514 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908555 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4deb0612-547f-4067-b95b-5794663d21aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908596 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:08 crc kubenswrapper[4810]: I1008 06:47:08.908616 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010245 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010300 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010339 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4deb0612-547f-4067-b95b-5794663d21aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010397 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010449 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010481 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010502 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4deb0612-547f-4067-b95b-5794663d21aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010530 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010562 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010590 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhz55\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-kube-api-access-zhz55\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.010652 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.011057 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.011136 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.012044 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.012399 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.013118 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.019517 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.019649 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4deb0612-547f-4067-b95b-5794663d21aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.023805 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4deb0612-547f-4067-b95b-5794663d21aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.024553 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.032166 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhz55\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-kube-api-access-zhz55\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.047260 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.146311 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.234932 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.622069 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2a44c96c-b96c-409c-9c9e-c049d9fe68b5","Type":"ContainerStarted","Data":"708b0993f45847c122345d0e8e83973a68b9edfdee190a0cbc4bc091346d9761"} Oct 08 06:47:09 crc kubenswrapper[4810]: I1008 06:47:09.771612 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 06:47:09 crc kubenswrapper[4810]: W1008 06:47:09.792764 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4deb0612_547f_4067_b95b_5794663d21aa.slice/crio-0bd79405f9b24e71b2a92fb1817c23cc334b58a5d026f67058ed9df15150c29d WatchSource:0}: Error finding container 0bd79405f9b24e71b2a92fb1817c23cc334b58a5d026f67058ed9df15150c29d: Status 404 returned error can't find the container with id 0bd79405f9b24e71b2a92fb1817c23cc334b58a5d026f67058ed9df15150c29d Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.363627 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.365023 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.368432 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.368700 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.369246 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.369371 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.371523 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gs7mv" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.377984 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.380441 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.464328 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.464740 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.464821 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-default\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.464848 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8cwm\" (UniqueName: \"kubernetes.io/projected/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kube-api-access-r8cwm\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.464912 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kolla-config\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.464982 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.465007 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-secrets\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.465033 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.465105 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566295 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566404 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566420 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566447 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-default\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8cwm\" (UniqueName: \"kubernetes.io/projected/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kube-api-access-r8cwm\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kolla-config\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566518 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.566537 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-secrets\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.568195 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.568774 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-default\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.569082 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.569324 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.572450 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kolla-config\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.573715 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-secrets\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.575164 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.590753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8cwm\" (UniqueName: \"kubernetes.io/projected/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kube-api-access-r8cwm\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.592486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.602880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " pod="openstack/openstack-galera-0" Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.648701 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4deb0612-547f-4067-b95b-5794663d21aa","Type":"ContainerStarted","Data":"0bd79405f9b24e71b2a92fb1817c23cc334b58a5d026f67058ed9df15150c29d"} Oct 08 06:47:10 crc kubenswrapper[4810]: I1008 06:47:10.713577 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.245513 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.509445 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.511164 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.514818 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.515015 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-kwqnj" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.515176 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.515370 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.539827 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591276 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591344 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591374 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591438 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nqm4\" (UniqueName: \"kubernetes.io/projected/a940c7a1-b3cc-4168-99e8-750c87845314-kube-api-access-4nqm4\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591617 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591873 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.591908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.592162 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.699707 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.699783 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.699836 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.699895 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.699956 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.701587 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.702940 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.703060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nqm4\" (UniqueName: \"kubernetes.io/projected/a940c7a1-b3cc-4168-99e8-750c87845314-kube-api-access-4nqm4\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.703098 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.703162 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.704433 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.705759 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.711783 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.711875 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.712289 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.712420 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.712524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.732927 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nqm4\" (UniqueName: \"kubernetes.io/projected/a940c7a1-b3cc-4168-99e8-750c87845314-kube-api-access-4nqm4\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.746842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.835463 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.906752 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.907798 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.912531 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.912590 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8njpq" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.912915 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 08 06:47:11 crc kubenswrapper[4810]: I1008 06:47:11.921728 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.024723 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-combined-ca-bundle\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.024814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kolla-config\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.024893 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ssjm\" (UniqueName: \"kubernetes.io/projected/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kube-api-access-7ssjm\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.025058 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-config-data\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.025124 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-memcached-tls-certs\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.128844 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-config-data\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.129679 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-config-data\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.129743 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-memcached-tls-certs\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.129787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-combined-ca-bundle\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.129824 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kolla-config\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.129857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ssjm\" (UniqueName: \"kubernetes.io/projected/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kube-api-access-7ssjm\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.132129 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kolla-config\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.136277 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-memcached-tls-certs\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.136523 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-combined-ca-bundle\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.150007 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ssjm\" (UniqueName: \"kubernetes.io/projected/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kube-api-access-7ssjm\") pod \"memcached-0\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " pod="openstack/memcached-0" Oct 08 06:47:12 crc kubenswrapper[4810]: I1008 06:47:12.298334 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.633339 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.635078 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.635391 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.639734 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vmmhr" Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.761938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq8jv\" (UniqueName: \"kubernetes.io/projected/372bbd1e-a3e6-4f87-b824-4ee000006789-kube-api-access-sq8jv\") pod \"kube-state-metrics-0\" (UID: \"372bbd1e-a3e6-4f87-b824-4ee000006789\") " pod="openstack/kube-state-metrics-0" Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.863655 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq8jv\" (UniqueName: \"kubernetes.io/projected/372bbd1e-a3e6-4f87-b824-4ee000006789-kube-api-access-sq8jv\") pod \"kube-state-metrics-0\" (UID: \"372bbd1e-a3e6-4f87-b824-4ee000006789\") " pod="openstack/kube-state-metrics-0" Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.907366 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq8jv\" (UniqueName: \"kubernetes.io/projected/372bbd1e-a3e6-4f87-b824-4ee000006789-kube-api-access-sq8jv\") pod \"kube-state-metrics-0\" (UID: \"372bbd1e-a3e6-4f87-b824-4ee000006789\") " pod="openstack/kube-state-metrics-0" Oct 08 06:47:13 crc kubenswrapper[4810]: I1008 06:47:13.966842 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:47:15 crc kubenswrapper[4810]: W1008 06:47:15.804653 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b86049c_4b80_4a1f_8e8b_51ad6e174fd7.slice/crio-706c6128ecbaea4077f4cb0e88470ac3c5e4ee2303ca3ba4fa21eb87439e4fc4 WatchSource:0}: Error finding container 706c6128ecbaea4077f4cb0e88470ac3c5e4ee2303ca3ba4fa21eb87439e4fc4: Status 404 returned error can't find the container with id 706c6128ecbaea4077f4cb0e88470ac3c5e4ee2303ca3ba4fa21eb87439e4fc4 Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.732079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7","Type":"ContainerStarted","Data":"706c6128ecbaea4077f4cb0e88470ac3c5e4ee2303ca3ba4fa21eb87439e4fc4"} Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.901875 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-qtj9x"] Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.902921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.906745 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rzcxq" Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.907455 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.910193 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.936020 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-xlsjm"] Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.938217 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.960404 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qtj9x"] Oct 08 06:47:16 crc kubenswrapper[4810]: I1008 06:47:16.972030 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xlsjm"] Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111360ca-3afe-4eb6-83a0-c60ca579c82b-scripts\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029483 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-ovn-controller-tls-certs\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029524 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run-ovn\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029565 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-combined-ca-bundle\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029594 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-log\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029616 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-lib\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029633 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hftz\" (UniqueName: \"kubernetes.io/projected/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-kube-api-access-6hftz\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029658 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-etc-ovs\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029675 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029725 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-log-ovn\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029747 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-scripts\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029766 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55m4\" (UniqueName: \"kubernetes.io/projected/111360ca-3afe-4eb6-83a0-c60ca579c82b-kube-api-access-x55m4\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.029785 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-run\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131382 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111360ca-3afe-4eb6-83a0-c60ca579c82b-scripts\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131439 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-ovn-controller-tls-certs\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131470 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run-ovn\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131506 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-combined-ca-bundle\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131530 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-log\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131555 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-lib\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hftz\" (UniqueName: \"kubernetes.io/projected/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-kube-api-access-6hftz\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131615 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-etc-ovs\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131683 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-log-ovn\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131707 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-scripts\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131726 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55m4\" (UniqueName: \"kubernetes.io/projected/111360ca-3afe-4eb6-83a0-c60ca579c82b-kube-api-access-x55m4\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.131746 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-run\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.132461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-run\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.132642 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-log\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.132781 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-lib\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.132898 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run-ovn\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.133130 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-log-ovn\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.133437 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-etc-ovs\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.133510 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.135802 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-scripts\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.139057 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111360ca-3afe-4eb6-83a0-c60ca579c82b-scripts\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.149680 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-combined-ca-bundle\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.155359 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-ovn-controller-tls-certs\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.158241 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55m4\" (UniqueName: \"kubernetes.io/projected/111360ca-3afe-4eb6-83a0-c60ca579c82b-kube-api-access-x55m4\") pod \"ovn-controller-qtj9x\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.179986 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hftz\" (UniqueName: \"kubernetes.io/projected/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-kube-api-access-6hftz\") pod \"ovn-controller-ovs-xlsjm\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.227805 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:17 crc kubenswrapper[4810]: I1008 06:47:17.266124 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.362032 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.365905 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.377368 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.377595 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.377708 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-b58vj" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.377896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.378094 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.378457 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.454526 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.454586 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgvsq\" (UniqueName: \"kubernetes.io/projected/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-kube-api-access-lgvsq\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.454711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-config\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.454792 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.454845 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.455011 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.455072 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.455227 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556240 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgvsq\" (UniqueName: \"kubernetes.io/projected/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-kube-api-access-lgvsq\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-config\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556329 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556365 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556417 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556440 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556478 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.556501 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.557120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.557603 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.558173 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.558351 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-config\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.563308 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.563662 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.575630 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.575883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgvsq\" (UniqueName: \"kubernetes.io/projected/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-kube-api-access-lgvsq\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.583497 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:18 crc kubenswrapper[4810]: I1008 06:47:18.699114 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.431662 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.432433 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.432537 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.435260 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf6102000b45e1894934eca5fb9ef2a49eefef58f2ba5b5554021b8f1388a9d9"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.435697 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://bf6102000b45e1894934eca5fb9ef2a49eefef58f2ba5b5554021b8f1388a9d9" gracePeriod=600 Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.679700 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.761936 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="bf6102000b45e1894934eca5fb9ef2a49eefef58f2ba5b5554021b8f1388a9d9" exitCode=0 Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.762040 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"bf6102000b45e1894934eca5fb9ef2a49eefef58f2ba5b5554021b8f1388a9d9"} Oct 08 06:47:19 crc kubenswrapper[4810]: I1008 06:47:19.762093 4810 scope.go:117] "RemoveContainer" containerID="2a91a7b94b569323b6b5755a125c6be2ca2114d957f75fbc9389bdbf60792809" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.222001 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.225630 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.232585 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.233129 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-d4x78" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.233518 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.234682 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.244612 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.329030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.329162 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mblm2\" (UniqueName: \"kubernetes.io/projected/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-kube-api-access-mblm2\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.329205 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.329233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.329290 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.329547 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.329629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.330066 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.432636 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mblm2\" (UniqueName: \"kubernetes.io/projected/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-kube-api-access-mblm2\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.432722 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.432765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.434138 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.434213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.434505 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.434536 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.434577 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.435446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.435811 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.436798 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.438063 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.444424 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.445168 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.446292 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.459075 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mblm2\" (UniqueName: \"kubernetes.io/projected/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-kube-api-access-mblm2\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.475800 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:21 crc kubenswrapper[4810]: I1008 06:47:21.560928 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:31 crc kubenswrapper[4810]: W1008 06:47:31.681847 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49cb68f6_1a48_4d79_90cf_34897b19dfaa.slice/crio-a49173aa83e289d210db693fd798ab7e9e14811ec7c1eecbd4f5b8356ed1f5bc WatchSource:0}: Error finding container a49173aa83e289d210db693fd798ab7e9e14811ec7c1eecbd4f5b8356ed1f5bc: Status 404 returned error can't find the container with id a49173aa83e289d210db693fd798ab7e9e14811ec7c1eecbd4f5b8356ed1f5bc Oct 08 06:47:31 crc kubenswrapper[4810]: I1008 06:47:31.786993 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 06:47:31 crc kubenswrapper[4810]: I1008 06:47:31.871465 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"49cb68f6-1a48-4d79-90cf-34897b19dfaa","Type":"ContainerStarted","Data":"a49173aa83e289d210db693fd798ab7e9e14811ec7c1eecbd4f5b8356ed1f5bc"} Oct 08 06:47:32 crc kubenswrapper[4810]: I1008 06:47:32.285387 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.792141 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.792606 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xjtgg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-77597f887-cq9qj_openstack(e046bc15-ade9-481c-a5c0-9c01f4794f36): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.796024 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-77597f887-cq9qj" podUID="e046bc15-ade9-481c-a5c0-9c01f4794f36" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.803383 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.803614 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xq5jg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-p4g8n_openstack(54cef57a-8119-4756-bea5-09f819e851b9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.817178 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" podUID="54cef57a-8119-4756-bea5-09f819e851b9" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.875975 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.876154 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rvsrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-644597f84c-krjcl_openstack(3e00f17c-b7d4-4b2d-80cb-81375f0c60ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.877642 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-644597f84c-krjcl" podUID="3e00f17c-b7d4-4b2d-80cb-81375f0c60ef" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.927799 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.928192 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6w2d6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-2j9fh_openstack(adaacb32-8c3c-4eb7-a376-a60927949bb9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.931291 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" podUID="adaacb32-8c3c-4eb7-a376-a60927949bb9" Oct 08 06:47:32 crc kubenswrapper[4810]: E1008 06:47:32.984876 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-77597f887-cq9qj" podUID="e046bc15-ade9-481c-a5c0-9c01f4794f36" Oct 08 06:47:33 crc kubenswrapper[4810]: W1008 06:47:33.052119 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda940c7a1_b3cc_4168_99e8_750c87845314.slice/crio-79eb39754856cde25b148052d7ca0631a6f8d5b24de4ca5267650373ae53d0e3 WatchSource:0}: Error finding container 79eb39754856cde25b148052d7ca0631a6f8d5b24de4ca5267650373ae53d0e3: Status 404 returned error can't find the container with id 79eb39754856cde25b148052d7ca0631a6f8d5b24de4ca5267650373ae53d0e3 Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.303766 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.398678 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-config\") pod \"54cef57a-8119-4756-bea5-09f819e851b9\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.398746 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-dns-svc\") pod \"54cef57a-8119-4756-bea5-09f819e851b9\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.398860 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq5jg\" (UniqueName: \"kubernetes.io/projected/54cef57a-8119-4756-bea5-09f819e851b9-kube-api-access-xq5jg\") pod \"54cef57a-8119-4756-bea5-09f819e851b9\" (UID: \"54cef57a-8119-4756-bea5-09f819e851b9\") " Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.400693 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-config" (OuterVolumeSpecName: "config") pod "54cef57a-8119-4756-bea5-09f819e851b9" (UID: "54cef57a-8119-4756-bea5-09f819e851b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.400722 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "54cef57a-8119-4756-bea5-09f819e851b9" (UID: "54cef57a-8119-4756-bea5-09f819e851b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.405080 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cef57a-8119-4756-bea5-09f819e851b9-kube-api-access-xq5jg" (OuterVolumeSpecName: "kube-api-access-xq5jg") pod "54cef57a-8119-4756-bea5-09f819e851b9" (UID: "54cef57a-8119-4756-bea5-09f819e851b9"). InnerVolumeSpecName "kube-api-access-xq5jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.504379 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq5jg\" (UniqueName: \"kubernetes.io/projected/54cef57a-8119-4756-bea5-09f819e851b9-kube-api-access-xq5jg\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.504425 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.504436 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cef57a-8119-4756-bea5-09f819e851b9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.789200 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qtj9x"] Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.853947 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.920103 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.934100 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7","Type":"ContainerStarted","Data":"1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b"} Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.940173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a9ba387c-c5ce-4f6e-8323-7e6afce894ba","Type":"ContainerStarted","Data":"916d488c5c8f7a36a08582b805ce1d1957f0e604d233e3dca47e1b5668fc8256"} Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.947622 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a940c7a1-b3cc-4168-99e8-750c87845314","Type":"ContainerStarted","Data":"85cbea479eda95f8290d55a3c5e21c38ca10eb0dbc85ea481f1c6dcd9a95fdc7"} Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.947691 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a940c7a1-b3cc-4168-99e8-750c87845314","Type":"ContainerStarted","Data":"79eb39754856cde25b148052d7ca0631a6f8d5b24de4ca5267650373ae53d0e3"} Oct 08 06:47:33 crc kubenswrapper[4810]: W1008 06:47:33.954446 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod372bbd1e_a3e6_4f87_b824_4ee000006789.slice/crio-c372558a4fa209c54d27868c155f6218c15707f4fd3c00ab75058d013de4576b WatchSource:0}: Error finding container c372558a4fa209c54d27868c155f6218c15707f4fd3c00ab75058d013de4576b: Status 404 returned error can't find the container with id c372558a4fa209c54d27868c155f6218c15707f4fd3c00ab75058d013de4576b Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.954872 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x" event={"ID":"111360ca-3afe-4eb6-83a0-c60ca579c82b","Type":"ContainerStarted","Data":"43d74d5e2fe2ea4b1d6bd07651d35dd091c2d05f4173f63194d7ff518c2a7bf3"} Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.961428 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.986876 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"fe33a8b708a76b97f4ea809aa4cdda57b3d8c666bcf720b830743227cfcf88de"} Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.992905 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" Oct 08 06:47:33 crc kubenswrapper[4810]: I1008 06:47:33.994027 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-p4g8n" event={"ID":"54cef57a-8119-4756-bea5-09f819e851b9","Type":"ContainerDied","Data":"071179ba28a7432d6bf5571c8d9940c38024601fb12f5024da982c4e43dd0d65"} Oct 08 06:47:34 crc kubenswrapper[4810]: E1008 06:47:34.006271 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-644597f84c-krjcl" podUID="3e00f17c-b7d4-4b2d-80cb-81375f0c60ef" Oct 08 06:47:34 crc kubenswrapper[4810]: W1008 06:47:34.064031 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a3bceca_c9a2_45e1_9f81_2bd9f6985c36.slice/crio-5919a28b07d4e5a1322abe30f98f1aba9bdeddb7cc6e3637d5ac8cb63ef9752d WatchSource:0}: Error finding container 5919a28b07d4e5a1322abe30f98f1aba9bdeddb7cc6e3637d5ac8cb63ef9752d: Status 404 returned error can't find the container with id 5919a28b07d4e5a1322abe30f98f1aba9bdeddb7cc6e3637d5ac8cb63ef9752d Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.169092 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-p4g8n"] Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.215786 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-p4g8n"] Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.465279 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.534714 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w2d6\" (UniqueName: \"kubernetes.io/projected/adaacb32-8c3c-4eb7-a376-a60927949bb9-kube-api-access-6w2d6\") pod \"adaacb32-8c3c-4eb7-a376-a60927949bb9\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.535055 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adaacb32-8c3c-4eb7-a376-a60927949bb9-config\") pod \"adaacb32-8c3c-4eb7-a376-a60927949bb9\" (UID: \"adaacb32-8c3c-4eb7-a376-a60927949bb9\") " Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.536381 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adaacb32-8c3c-4eb7-a376-a60927949bb9-config" (OuterVolumeSpecName: "config") pod "adaacb32-8c3c-4eb7-a376-a60927949bb9" (UID: "adaacb32-8c3c-4eb7-a376-a60927949bb9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.545127 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adaacb32-8c3c-4eb7-a376-a60927949bb9-kube-api-access-6w2d6" (OuterVolumeSpecName: "kube-api-access-6w2d6") pod "adaacb32-8c3c-4eb7-a376-a60927949bb9" (UID: "adaacb32-8c3c-4eb7-a376-a60927949bb9"). InnerVolumeSpecName "kube-api-access-6w2d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.637615 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adaacb32-8c3c-4eb7-a376-a60927949bb9-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.637681 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w2d6\" (UniqueName: \"kubernetes.io/projected/adaacb32-8c3c-4eb7-a376-a60927949bb9-kube-api-access-6w2d6\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:34 crc kubenswrapper[4810]: I1008 06:47:34.869721 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xlsjm"] Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.021148 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"372bbd1e-a3e6-4f87-b824-4ee000006789","Type":"ContainerStarted","Data":"c372558a4fa209c54d27868c155f6218c15707f4fd3c00ab75058d013de4576b"} Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.025559 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" event={"ID":"adaacb32-8c3c-4eb7-a376-a60927949bb9","Type":"ContainerDied","Data":"0f0fe5d95e6c526565b0d2f7a06adb42781ba0b2091b7da5919648e8502934ac"} Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.025704 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-2j9fh" Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.031853 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36","Type":"ContainerStarted","Data":"5919a28b07d4e5a1322abe30f98f1aba9bdeddb7cc6e3637d5ac8cb63ef9752d"} Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.045088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2a44c96c-b96c-409c-9c9e-c049d9fe68b5","Type":"ContainerStarted","Data":"a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1"} Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.059668 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4deb0612-547f-4067-b95b-5794663d21aa","Type":"ContainerStarted","Data":"c3755a82f85d0942937435d7ed4103b19749825c63d0c7188863bb653199761b"} Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.213162 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-2j9fh"] Oct 08 06:47:35 crc kubenswrapper[4810]: I1008 06:47:35.228641 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-2j9fh"] Oct 08 06:47:35 crc kubenswrapper[4810]: W1008 06:47:35.716290 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc5c4854_7fbb_4810_9dfc_cc22d6cdbd37.slice/crio-95eab66a50345cf0bb53a0a140c2990323fab8df4865001953b83fd3efd14cd9 WatchSource:0}: Error finding container 95eab66a50345cf0bb53a0a140c2990323fab8df4865001953b83fd3efd14cd9: Status 404 returned error can't find the container with id 95eab66a50345cf0bb53a0a140c2990323fab8df4865001953b83fd3efd14cd9 Oct 08 06:47:36 crc kubenswrapper[4810]: I1008 06:47:36.085347 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54cef57a-8119-4756-bea5-09f819e851b9" path="/var/lib/kubelet/pods/54cef57a-8119-4756-bea5-09f819e851b9/volumes" Oct 08 06:47:36 crc kubenswrapper[4810]: I1008 06:47:36.085901 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adaacb32-8c3c-4eb7-a376-a60927949bb9" path="/var/lib/kubelet/pods/adaacb32-8c3c-4eb7-a376-a60927949bb9/volumes" Oct 08 06:47:36 crc kubenswrapper[4810]: I1008 06:47:36.086444 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xlsjm" event={"ID":"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37","Type":"ContainerStarted","Data":"95eab66a50345cf0bb53a0a140c2990323fab8df4865001953b83fd3efd14cd9"} Oct 08 06:47:38 crc kubenswrapper[4810]: I1008 06:47:38.104547 4810 generic.go:334] "Generic (PLEG): container finished" podID="a940c7a1-b3cc-4168-99e8-750c87845314" containerID="85cbea479eda95f8290d55a3c5e21c38ca10eb0dbc85ea481f1c6dcd9a95fdc7" exitCode=0 Oct 08 06:47:38 crc kubenswrapper[4810]: I1008 06:47:38.104641 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a940c7a1-b3cc-4168-99e8-750c87845314","Type":"ContainerDied","Data":"85cbea479eda95f8290d55a3c5e21c38ca10eb0dbc85ea481f1c6dcd9a95fdc7"} Oct 08 06:47:38 crc kubenswrapper[4810]: I1008 06:47:38.108351 4810 generic.go:334] "Generic (PLEG): container finished" podID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerID="1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b" exitCode=0 Oct 08 06:47:38 crc kubenswrapper[4810]: I1008 06:47:38.108425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7","Type":"ContainerDied","Data":"1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.138768 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7","Type":"ContainerStarted","Data":"70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.143878 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"372bbd1e-a3e6-4f87-b824-4ee000006789","Type":"ContainerStarted","Data":"d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.144082 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.145822 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"49cb68f6-1a48-4d79-90cf-34897b19dfaa","Type":"ContainerStarted","Data":"9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.146172 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.147667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36","Type":"ContainerStarted","Data":"e4123f776f7b7b29b557fdc6e108f0d0848f10b0a0ef5e629afe696812883bbf"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.149694 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a9ba387c-c5ce-4f6e-8323-7e6afce894ba","Type":"ContainerStarted","Data":"9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.152224 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a940c7a1-b3cc-4168-99e8-750c87845314","Type":"ContainerStarted","Data":"dd78d50a7f1c3827da6046339729794dc53b8121adbd9c479e00e2f78cbf711a"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.153816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x" event={"ID":"111360ca-3afe-4eb6-83a0-c60ca579c82b","Type":"ContainerStarted","Data":"3110ad491e155ae79f17fbaedfb3c85a0be30ad15f2ba193fc7eb67f4b9b78a1"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.154023 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-qtj9x" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.156070 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerID="5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537" exitCode=0 Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.156182 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xlsjm" event={"ID":"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37","Type":"ContainerDied","Data":"5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537"} Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.172027 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=14.90646682 podStartE2EDuration="32.172002164s" podCreationTimestamp="2025-10-08 06:47:09 +0000 UTC" firstStartedPulling="2025-10-08 06:47:15.820261845 +0000 UTC m=+938.454701595" lastFinishedPulling="2025-10-08 06:47:33.085797199 +0000 UTC m=+955.720236939" observedRunningTime="2025-10-08 06:47:41.162841243 +0000 UTC m=+963.797280983" watchObservedRunningTime="2025-10-08 06:47:41.172002164 +0000 UTC m=+963.806441904" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.197670 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=31.197649407 podStartE2EDuration="31.197649407s" podCreationTimestamp="2025-10-08 06:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:47:41.192349581 +0000 UTC m=+963.826789321" watchObservedRunningTime="2025-10-08 06:47:41.197649407 +0000 UTC m=+963.832089147" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.239570 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=22.259346689 podStartE2EDuration="28.239541085s" podCreationTimestamp="2025-10-08 06:47:13 +0000 UTC" firstStartedPulling="2025-10-08 06:47:33.963687223 +0000 UTC m=+956.598126963" lastFinishedPulling="2025-10-08 06:47:39.943881619 +0000 UTC m=+962.578321359" observedRunningTime="2025-10-08 06:47:41.21127349 +0000 UTC m=+963.845713230" watchObservedRunningTime="2025-10-08 06:47:41.239541085 +0000 UTC m=+963.873980835" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.270733 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=23.320041281 podStartE2EDuration="30.270708499s" podCreationTimestamp="2025-10-08 06:47:11 +0000 UTC" firstStartedPulling="2025-10-08 06:47:31.786414471 +0000 UTC m=+954.420854251" lastFinishedPulling="2025-10-08 06:47:38.737081689 +0000 UTC m=+961.371521469" observedRunningTime="2025-10-08 06:47:41.257866958 +0000 UTC m=+963.892306708" watchObservedRunningTime="2025-10-08 06:47:41.270708499 +0000 UTC m=+963.905148229" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.287771 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-qtj9x" podStartSLOduration=19.951123731 podStartE2EDuration="25.287745786s" podCreationTimestamp="2025-10-08 06:47:16 +0000 UTC" firstStartedPulling="2025-10-08 06:47:33.814588156 +0000 UTC m=+956.449027896" lastFinishedPulling="2025-10-08 06:47:39.151210211 +0000 UTC m=+961.785649951" observedRunningTime="2025-10-08 06:47:41.284949409 +0000 UTC m=+963.919389149" watchObservedRunningTime="2025-10-08 06:47:41.287745786 +0000 UTC m=+963.922185526" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.837420 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:41 crc kubenswrapper[4810]: I1008 06:47:41.838092 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:42 crc kubenswrapper[4810]: I1008 06:47:42.169012 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xlsjm" event={"ID":"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37","Type":"ContainerStarted","Data":"785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611"} Oct 08 06:47:42 crc kubenswrapper[4810]: I1008 06:47:42.169109 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xlsjm" event={"ID":"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37","Type":"ContainerStarted","Data":"f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775"} Oct 08 06:47:42 crc kubenswrapper[4810]: I1008 06:47:42.197112 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-xlsjm" podStartSLOduration=22.649995692 podStartE2EDuration="26.197089703s" podCreationTimestamp="2025-10-08 06:47:16 +0000 UTC" firstStartedPulling="2025-10-08 06:47:35.720114009 +0000 UTC m=+958.354553749" lastFinishedPulling="2025-10-08 06:47:39.267208 +0000 UTC m=+961.901647760" observedRunningTime="2025-10-08 06:47:42.19479112 +0000 UTC m=+964.829230860" watchObservedRunningTime="2025-10-08 06:47:42.197089703 +0000 UTC m=+964.831529443" Oct 08 06:47:42 crc kubenswrapper[4810]: I1008 06:47:42.266605 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:42 crc kubenswrapper[4810]: I1008 06:47:42.266846 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:47:44 crc kubenswrapper[4810]: I1008 06:47:44.189235 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36","Type":"ContainerStarted","Data":"02937c8f188e832e3f815057f823449d155b5ae31271a2816bdd2a5efb1d99a8"} Oct 08 06:47:44 crc kubenswrapper[4810]: I1008 06:47:44.190732 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a9ba387c-c5ce-4f6e-8323-7e6afce894ba","Type":"ContainerStarted","Data":"b2cbef70cc311c4f05f44991c29903743960b2ccd17d6f94057bd5df3a186bbb"} Oct 08 06:47:44 crc kubenswrapper[4810]: I1008 06:47:44.223259 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.612600922 podStartE2EDuration="27.223237383s" podCreationTimestamp="2025-10-08 06:47:17 +0000 UTC" firstStartedPulling="2025-10-08 06:47:34.069058422 +0000 UTC m=+956.703498162" lastFinishedPulling="2025-10-08 06:47:43.679694883 +0000 UTC m=+966.314134623" observedRunningTime="2025-10-08 06:47:44.214867193 +0000 UTC m=+966.849306973" watchObservedRunningTime="2025-10-08 06:47:44.223237383 +0000 UTC m=+966.857677123" Oct 08 06:47:44 crc kubenswrapper[4810]: I1008 06:47:44.251590 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=14.453613971 podStartE2EDuration="24.251569659s" podCreationTimestamp="2025-10-08 06:47:20 +0000 UTC" firstStartedPulling="2025-10-08 06:47:33.895068721 +0000 UTC m=+956.529508461" lastFinishedPulling="2025-10-08 06:47:43.693024409 +0000 UTC m=+966.327464149" observedRunningTime="2025-10-08 06:47:44.248825894 +0000 UTC m=+966.883265684" watchObservedRunningTime="2025-10-08 06:47:44.251569659 +0000 UTC m=+966.886009399" Oct 08 06:47:45 crc kubenswrapper[4810]: I1008 06:47:45.561460 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:45 crc kubenswrapper[4810]: I1008 06:47:45.642796 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:45 crc kubenswrapper[4810]: I1008 06:47:45.699885 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:45 crc kubenswrapper[4810]: I1008 06:47:45.759441 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.210447 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.210532 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.268234 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.276628 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.574519 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-krjcl"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.637185 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-d6hjw"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.639478 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.643396 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.653576 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-fchxc"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.654736 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.658558 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.693856 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-d6hjw"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.694988 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms7w4\" (UniqueName: \"kubernetes.io/projected/22e37514-b509-4b91-b3c5-7a30006a4322-kube-api-access-ms7w4\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695035 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-config\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695078 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f5fbdd-0691-4ce8-888d-be0147974607-config\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695098 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc2pw\" (UniqueName: \"kubernetes.io/projected/d6f5fbdd-0691-4ce8-888d-be0147974607-kube-api-access-qc2pw\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695145 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-ovsdbserver-sb\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695163 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovn-rundir\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-dns-svc\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695259 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-combined-ca-bundle\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.695275 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovs-rundir\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.728070 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fchxc"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.778043 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-cq9qj"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.781089 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.785224 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.796580 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.796760 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.796874 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.797094 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-rbfpg" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.797937 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-ovsdbserver-sb\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.797981 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovn-rundir\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.798011 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.798037 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-dns-svc\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.798787 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-ovsdbserver-sb\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.799024 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovn-rundir\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.799914 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-dns-svc\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.803478 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-combined-ca-bundle\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.803540 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovs-rundir\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.803776 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms7w4\" (UniqueName: \"kubernetes.io/projected/22e37514-b509-4b91-b3c5-7a30006a4322-kube-api-access-ms7w4\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.803832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-config\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.803875 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f5fbdd-0691-4ce8-888d-be0147974607-config\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.803903 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc2pw\" (UniqueName: \"kubernetes.io/projected/d6f5fbdd-0691-4ce8-888d-be0147974607-kube-api-access-qc2pw\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.804876 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.805283 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-config\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.805493 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovs-rundir\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.805798 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f5fbdd-0691-4ce8-888d-be0147974607-config\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.828716 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-combined-ca-bundle\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.835287 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-nmfc4"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.836684 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.838646 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.851420 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc2pw\" (UniqueName: \"kubernetes.io/projected/d6f5fbdd-0691-4ce8-888d-be0147974607-kube-api-access-qc2pw\") pod \"ovn-controller-metrics-fchxc\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.851690 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.872654 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms7w4\" (UniqueName: \"kubernetes.io/projected/22e37514-b509-4b91-b3c5-7a30006a4322-kube-api-access-ms7w4\") pod \"dnsmasq-dns-545fb8c44f-d6hjw\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910395 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bprw5\" (UniqueName: \"kubernetes.io/projected/df9f8bfe-4f8d-4f95-910c-0065aa227782-kube-api-access-bprw5\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910458 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-scripts\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910491 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910567 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910601 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910650 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910760 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-config\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910790 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910819 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-config\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910858 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp7gp\" (UniqueName: \"kubernetes.io/projected/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-kube-api-access-mp7gp\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910889 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.910951 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.919068 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-nmfc4"] Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.979155 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:46 crc kubenswrapper[4810]: I1008 06:47:46.995397 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015152 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-config\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015177 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015197 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-config\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015222 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp7gp\" (UniqueName: \"kubernetes.io/projected/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-kube-api-access-mp7gp\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015268 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015309 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bprw5\" (UniqueName: \"kubernetes.io/projected/df9f8bfe-4f8d-4f95-910c-0065aa227782-kube-api-access-bprw5\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015332 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-scripts\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015353 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015396 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.015413 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.016464 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.017753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.017893 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-scripts\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.018325 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-config\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.021111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-config\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.021292 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.021417 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.024572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.025862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.028512 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.057054 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp7gp\" (UniqueName: \"kubernetes.io/projected/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-kube-api-access-mp7gp\") pod \"ovn-northd-0\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.057675 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bprw5\" (UniqueName: \"kubernetes.io/projected/df9f8bfe-4f8d-4f95-910c-0065aa227782-kube-api-access-bprw5\") pod \"dnsmasq-dns-dc9d58d7-nmfc4\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.068376 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.115982 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-config\") pod \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.116050 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvsrd\" (UniqueName: \"kubernetes.io/projected/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-kube-api-access-rvsrd\") pod \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.116187 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-dns-svc\") pod \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\" (UID: \"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef\") " Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.117381 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-config" (OuterVolumeSpecName: "config") pod "3e00f17c-b7d4-4b2d-80cb-81375f0c60ef" (UID: "3e00f17c-b7d4-4b2d-80cb-81375f0c60ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.117600 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e00f17c-b7d4-4b2d-80cb-81375f0c60ef" (UID: "3e00f17c-b7d4-4b2d-80cb-81375f0c60ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.120174 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-kube-api-access-rvsrd" (OuterVolumeSpecName: "kube-api-access-rvsrd") pod "3e00f17c-b7d4-4b2d-80cb-81375f0c60ef" (UID: "3e00f17c-b7d4-4b2d-80cb-81375f0c60ef"). InnerVolumeSpecName "kube-api-access-rvsrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.205758 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.216372 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-krjcl" event={"ID":"3e00f17c-b7d4-4b2d-80cb-81375f0c60ef","Type":"ContainerDied","Data":"65d1515a7ea2337a45c74c7a240ee260cf60f3c6b3077371578a2b385addb848"} Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.216450 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-krjcl" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.218273 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjtgg\" (UniqueName: \"kubernetes.io/projected/e046bc15-ade9-481c-a5c0-9c01f4794f36-kube-api-access-xjtgg\") pod \"e046bc15-ade9-481c-a5c0-9c01f4794f36\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.218311 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-dns-svc\") pod \"e046bc15-ade9-481c-a5c0-9c01f4794f36\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.218331 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-config\") pod \"e046bc15-ade9-481c-a5c0-9c01f4794f36\" (UID: \"e046bc15-ade9-481c-a5c0-9c01f4794f36\") " Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.219318 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e046bc15-ade9-481c-a5c0-9c01f4794f36" (UID: "e046bc15-ade9-481c-a5c0-9c01f4794f36"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.219334 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-config" (OuterVolumeSpecName: "config") pod "e046bc15-ade9-481c-a5c0-9c01f4794f36" (UID: "e046bc15-ade9-481c-a5c0-9c01f4794f36"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.219709 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.219722 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.219731 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e046bc15-ade9-481c-a5c0-9c01f4794f36-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.219741 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvsrd\" (UniqueName: \"kubernetes.io/projected/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-kube-api-access-rvsrd\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.219750 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.220939 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-cq9qj" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.221095 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-cq9qj" event={"ID":"e046bc15-ade9-481c-a5c0-9c01f4794f36","Type":"ContainerDied","Data":"98191b8b4060a38247915b86bc963821f84618c63106ba35a6b155009ab40f43"} Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.222002 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e046bc15-ade9-481c-a5c0-9c01f4794f36-kube-api-access-xjtgg" (OuterVolumeSpecName: "kube-api-access-xjtgg") pod "e046bc15-ade9-481c-a5c0-9c01f4794f36" (UID: "e046bc15-ade9-481c-a5c0-9c01f4794f36"). InnerVolumeSpecName "kube-api-access-xjtgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.257385 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.279280 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.299545 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-krjcl"] Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.301589 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.314062 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-krjcl"] Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.322567 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjtgg\" (UniqueName: \"kubernetes.io/projected/e046bc15-ade9-481c-a5c0-9c01f4794f36-kube-api-access-xjtgg\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.380526 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-d6hjw"] Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.435332 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fchxc"] Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.600076 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-cq9qj"] Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.605052 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-cq9qj"] Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.665175 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-nmfc4"] Oct 08 06:47:47 crc kubenswrapper[4810]: I1008 06:47:47.841001 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.088804 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e00f17c-b7d4-4b2d-80cb-81375f0c60ef" path="/var/lib/kubelet/pods/3e00f17c-b7d4-4b2d-80cb-81375f0c60ef/volumes" Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.089819 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e046bc15-ade9-481c-a5c0-9c01f4794f36" path="/var/lib/kubelet/pods/e046bc15-ade9-481c-a5c0-9c01f4794f36/volumes" Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.236138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" event={"ID":"22e37514-b509-4b91-b3c5-7a30006a4322","Type":"ContainerDied","Data":"f9650e8f4bf603d068c24d042efbc2d4b597e5c795159be5d3e9daed83ee6a6d"} Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.235669 4810 generic.go:334] "Generic (PLEG): container finished" podID="22e37514-b509-4b91-b3c5-7a30006a4322" containerID="f9650e8f4bf603d068c24d042efbc2d4b597e5c795159be5d3e9daed83ee6a6d" exitCode=0 Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.236835 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" event={"ID":"22e37514-b509-4b91-b3c5-7a30006a4322","Type":"ContainerStarted","Data":"4617bea7d2120f432a392fbf6dd130c534149fc427d8d5bb0ee0833b2acd145f"} Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.243400 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fchxc" event={"ID":"d6f5fbdd-0691-4ce8-888d-be0147974607","Type":"ContainerStarted","Data":"8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49"} Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.243758 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fchxc" event={"ID":"d6f5fbdd-0691-4ce8-888d-be0147974607","Type":"ContainerStarted","Data":"2c85a732a55ba049badae5cd7192c567574150bda1f0d37daccf3ad3ec2b4834"} Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.245244 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" event={"ID":"df9f8bfe-4f8d-4f95-910c-0065aa227782","Type":"ContainerStarted","Data":"08ef1d016c592ff78e29f02e93e848b41352d6fc78b6b60d1175cf73859a704d"} Oct 08 06:47:48 crc kubenswrapper[4810]: I1008 06:47:48.263043 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7","Type":"ContainerStarted","Data":"ee3158257410582bd8eae3a8191820076a2b25b91711a64ef3ac1ac03ae953d1"} Oct 08 06:47:49 crc kubenswrapper[4810]: I1008 06:47:49.285954 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" event={"ID":"22e37514-b509-4b91-b3c5-7a30006a4322","Type":"ContainerStarted","Data":"61b720b6445fc80efcd5568efd1f845e0db4dbd75a26d415920ab472a986d592"} Oct 08 06:47:49 crc kubenswrapper[4810]: I1008 06:47:49.286521 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:49 crc kubenswrapper[4810]: I1008 06:47:49.287928 4810 generic.go:334] "Generic (PLEG): container finished" podID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerID="9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b" exitCode=0 Oct 08 06:47:49 crc kubenswrapper[4810]: I1008 06:47:49.288042 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" event={"ID":"df9f8bfe-4f8d-4f95-910c-0065aa227782","Type":"ContainerDied","Data":"9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b"} Oct 08 06:47:49 crc kubenswrapper[4810]: I1008 06:47:49.308040 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" podStartSLOduration=2.8535370970000002 podStartE2EDuration="3.308012454s" podCreationTimestamp="2025-10-08 06:47:46 +0000 UTC" firstStartedPulling="2025-10-08 06:47:47.41083487 +0000 UTC m=+970.045274610" lastFinishedPulling="2025-10-08 06:47:47.865310217 +0000 UTC m=+970.499749967" observedRunningTime="2025-10-08 06:47:49.306637057 +0000 UTC m=+971.941076797" watchObservedRunningTime="2025-10-08 06:47:49.308012454 +0000 UTC m=+971.942452194" Oct 08 06:47:49 crc kubenswrapper[4810]: I1008 06:47:49.313640 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-fchxc" podStartSLOduration=3.313613147 podStartE2EDuration="3.313613147s" podCreationTimestamp="2025-10-08 06:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:47:48.280116668 +0000 UTC m=+970.914556418" watchObservedRunningTime="2025-10-08 06:47:49.313613147 +0000 UTC m=+971.948052887" Oct 08 06:47:49 crc kubenswrapper[4810]: I1008 06:47:49.958943 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.014403 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.299836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" event={"ID":"df9f8bfe-4f8d-4f95-910c-0065aa227782","Type":"ContainerStarted","Data":"665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea"} Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.299936 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.302882 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7","Type":"ContainerStarted","Data":"6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b"} Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.302927 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7","Type":"ContainerStarted","Data":"a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c"} Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.303146 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.327580 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" podStartSLOduration=3.777416818 podStartE2EDuration="4.327564768s" podCreationTimestamp="2025-10-08 06:47:46 +0000 UTC" firstStartedPulling="2025-10-08 06:47:47.701518068 +0000 UTC m=+970.335957808" lastFinishedPulling="2025-10-08 06:47:48.251666018 +0000 UTC m=+970.886105758" observedRunningTime="2025-10-08 06:47:50.318151547 +0000 UTC m=+972.952591297" watchObservedRunningTime="2025-10-08 06:47:50.327564768 +0000 UTC m=+972.962004508" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.346013 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.991480511 podStartE2EDuration="4.34599542s" podCreationTimestamp="2025-10-08 06:47:46 +0000 UTC" firstStartedPulling="2025-10-08 06:47:47.868567067 +0000 UTC m=+970.503006807" lastFinishedPulling="2025-10-08 06:47:49.223081976 +0000 UTC m=+971.857521716" observedRunningTime="2025-10-08 06:47:50.338954524 +0000 UTC m=+972.973394254" watchObservedRunningTime="2025-10-08 06:47:50.34599542 +0000 UTC m=+972.980435160" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.713933 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.714040 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 08 06:47:50 crc kubenswrapper[4810]: I1008 06:47:50.768642 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 08 06:47:51 crc kubenswrapper[4810]: I1008 06:47:51.388940 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 08 06:47:51 crc kubenswrapper[4810]: I1008 06:47:51.939535 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-kfjh7"] Oct 08 06:47:51 crc kubenswrapper[4810]: I1008 06:47:51.943810 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfjh7" Oct 08 06:47:51 crc kubenswrapper[4810]: I1008 06:47:51.957938 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kfjh7"] Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.047436 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d95bq\" (UniqueName: \"kubernetes.io/projected/6dce6778-383f-4471-b852-16eaed68e455-kube-api-access-d95bq\") pod \"keystone-db-create-kfjh7\" (UID: \"6dce6778-383f-4471-b852-16eaed68e455\") " pod="openstack/keystone-db-create-kfjh7" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.149505 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d95bq\" (UniqueName: \"kubernetes.io/projected/6dce6778-383f-4471-b852-16eaed68e455-kube-api-access-d95bq\") pod \"keystone-db-create-kfjh7\" (UID: \"6dce6778-383f-4471-b852-16eaed68e455\") " pod="openstack/keystone-db-create-kfjh7" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.179233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d95bq\" (UniqueName: \"kubernetes.io/projected/6dce6778-383f-4471-b852-16eaed68e455-kube-api-access-d95bq\") pod \"keystone-db-create-kfjh7\" (UID: \"6dce6778-383f-4471-b852-16eaed68e455\") " pod="openstack/keystone-db-create-kfjh7" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.260219 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-hx4fs"] Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.262032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hx4fs" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.270991 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfjh7" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.282439 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hx4fs"] Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.352935 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tmw2\" (UniqueName: \"kubernetes.io/projected/508940e7-94fa-4ada-a7b6-0082a5fa6091-kube-api-access-4tmw2\") pod \"placement-db-create-hx4fs\" (UID: \"508940e7-94fa-4ada-a7b6-0082a5fa6091\") " pod="openstack/placement-db-create-hx4fs" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.454302 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tmw2\" (UniqueName: \"kubernetes.io/projected/508940e7-94fa-4ada-a7b6-0082a5fa6091-kube-api-access-4tmw2\") pod \"placement-db-create-hx4fs\" (UID: \"508940e7-94fa-4ada-a7b6-0082a5fa6091\") " pod="openstack/placement-db-create-hx4fs" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.483923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tmw2\" (UniqueName: \"kubernetes.io/projected/508940e7-94fa-4ada-a7b6-0082a5fa6091-kube-api-access-4tmw2\") pod \"placement-db-create-hx4fs\" (UID: \"508940e7-94fa-4ada-a7b6-0082a5fa6091\") " pod="openstack/placement-db-create-hx4fs" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.596894 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hx4fs" Oct 08 06:47:52 crc kubenswrapper[4810]: I1008 06:47:52.745180 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kfjh7"] Oct 08 06:47:52 crc kubenswrapper[4810]: W1008 06:47:52.773585 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dce6778_383f_4471_b852_16eaed68e455.slice/crio-709c082754f29d82d3099868563c0442d68e2646c540edd82f5309992958b1f2 WatchSource:0}: Error finding container 709c082754f29d82d3099868563c0442d68e2646c540edd82f5309992958b1f2: Status 404 returned error can't find the container with id 709c082754f29d82d3099868563c0442d68e2646c540edd82f5309992958b1f2 Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.039371 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hx4fs"] Oct 08 06:47:53 crc kubenswrapper[4810]: W1008 06:47:53.060632 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod508940e7_94fa_4ada_a7b6_0082a5fa6091.slice/crio-b434e9828be4313500927eeee24f26abae2d8f725b1f15bbe5dbaa195e3d1fca WatchSource:0}: Error finding container b434e9828be4313500927eeee24f26abae2d8f725b1f15bbe5dbaa195e3d1fca: Status 404 returned error can't find the container with id b434e9828be4313500927eeee24f26abae2d8f725b1f15bbe5dbaa195e3d1fca Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.342271 4810 generic.go:334] "Generic (PLEG): container finished" podID="6dce6778-383f-4471-b852-16eaed68e455" containerID="21bf0b67108876603e6a3ca6e2697321f8c85f0591ff8f0d50ebf6f57df7b7b3" exitCode=0 Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.342324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kfjh7" event={"ID":"6dce6778-383f-4471-b852-16eaed68e455","Type":"ContainerDied","Data":"21bf0b67108876603e6a3ca6e2697321f8c85f0591ff8f0d50ebf6f57df7b7b3"} Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.342681 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kfjh7" event={"ID":"6dce6778-383f-4471-b852-16eaed68e455","Type":"ContainerStarted","Data":"709c082754f29d82d3099868563c0442d68e2646c540edd82f5309992958b1f2"} Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.344444 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hx4fs" event={"ID":"508940e7-94fa-4ada-a7b6-0082a5fa6091","Type":"ContainerStarted","Data":"5a780ab4f9043af07e2e7ba563e150db2ea7db7ad6e6b93d0b0d6250802624cd"} Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.344938 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hx4fs" event={"ID":"508940e7-94fa-4ada-a7b6-0082a5fa6091","Type":"ContainerStarted","Data":"b434e9828be4313500927eeee24f26abae2d8f725b1f15bbe5dbaa195e3d1fca"} Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.432157 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-hx4fs" podStartSLOduration=1.432136768 podStartE2EDuration="1.432136768s" podCreationTimestamp="2025-10-08 06:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:47:53.427978772 +0000 UTC m=+976.062418512" watchObservedRunningTime="2025-10-08 06:47:53.432136768 +0000 UTC m=+976.066576508" Oct 08 06:47:53 crc kubenswrapper[4810]: I1008 06:47:53.997841 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.031876 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-d6hjw"] Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.032314 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" podUID="22e37514-b509-4b91-b3c5-7a30006a4322" containerName="dnsmasq-dns" containerID="cri-o://61b720b6445fc80efcd5568efd1f845e0db4dbd75a26d415920ab472a986d592" gracePeriod=10 Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.033147 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.064220 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-xqcwk"] Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.065547 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.154621 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-xqcwk"] Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.190867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.190937 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8xvn\" (UniqueName: \"kubernetes.io/projected/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-kube-api-access-p8xvn\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.190986 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-config\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.191035 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.191100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-dns-svc\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.292262 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-dns-svc\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.292413 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.292450 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8xvn\" (UniqueName: \"kubernetes.io/projected/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-kube-api-access-p8xvn\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.292472 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-config\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.292502 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.293434 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.293438 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-dns-svc\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.294062 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-config\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.294098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.325551 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8xvn\" (UniqueName: \"kubernetes.io/projected/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-kube-api-access-p8xvn\") pod \"dnsmasq-dns-7b587f8db7-xqcwk\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.361940 4810 generic.go:334] "Generic (PLEG): container finished" podID="22e37514-b509-4b91-b3c5-7a30006a4322" containerID="61b720b6445fc80efcd5568efd1f845e0db4dbd75a26d415920ab472a986d592" exitCode=0 Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.362151 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" event={"ID":"22e37514-b509-4b91-b3c5-7a30006a4322","Type":"ContainerDied","Data":"61b720b6445fc80efcd5568efd1f845e0db4dbd75a26d415920ab472a986d592"} Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.371818 4810 generic.go:334] "Generic (PLEG): container finished" podID="508940e7-94fa-4ada-a7b6-0082a5fa6091" containerID="5a780ab4f9043af07e2e7ba563e150db2ea7db7ad6e6b93d0b0d6250802624cd" exitCode=0 Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.372166 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hx4fs" event={"ID":"508940e7-94fa-4ada-a7b6-0082a5fa6091","Type":"ContainerDied","Data":"5a780ab4f9043af07e2e7ba563e150db2ea7db7ad6e6b93d0b0d6250802624cd"} Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.430755 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.583373 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.698809 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-ovsdbserver-sb\") pod \"22e37514-b509-4b91-b3c5-7a30006a4322\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.699590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms7w4\" (UniqueName: \"kubernetes.io/projected/22e37514-b509-4b91-b3c5-7a30006a4322-kube-api-access-ms7w4\") pod \"22e37514-b509-4b91-b3c5-7a30006a4322\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.699644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-config\") pod \"22e37514-b509-4b91-b3c5-7a30006a4322\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.699784 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-dns-svc\") pod \"22e37514-b509-4b91-b3c5-7a30006a4322\" (UID: \"22e37514-b509-4b91-b3c5-7a30006a4322\") " Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.711941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22e37514-b509-4b91-b3c5-7a30006a4322-kube-api-access-ms7w4" (OuterVolumeSpecName: "kube-api-access-ms7w4") pod "22e37514-b509-4b91-b3c5-7a30006a4322" (UID: "22e37514-b509-4b91-b3c5-7a30006a4322"). InnerVolumeSpecName "kube-api-access-ms7w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.757125 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "22e37514-b509-4b91-b3c5-7a30006a4322" (UID: "22e37514-b509-4b91-b3c5-7a30006a4322"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.764468 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "22e37514-b509-4b91-b3c5-7a30006a4322" (UID: "22e37514-b509-4b91-b3c5-7a30006a4322"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.781526 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-config" (OuterVolumeSpecName: "config") pod "22e37514-b509-4b91-b3c5-7a30006a4322" (UID: "22e37514-b509-4b91-b3c5-7a30006a4322"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.798875 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfjh7" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.801735 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.801763 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.801775 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms7w4\" (UniqueName: \"kubernetes.io/projected/22e37514-b509-4b91-b3c5-7a30006a4322-kube-api-access-ms7w4\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.801787 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e37514-b509-4b91-b3c5-7a30006a4322-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.903401 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d95bq\" (UniqueName: \"kubernetes.io/projected/6dce6778-383f-4471-b852-16eaed68e455-kube-api-access-d95bq\") pod \"6dce6778-383f-4471-b852-16eaed68e455\" (UID: \"6dce6778-383f-4471-b852-16eaed68e455\") " Oct 08 06:47:54 crc kubenswrapper[4810]: I1008 06:47:54.907364 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dce6778-383f-4471-b852-16eaed68e455-kube-api-access-d95bq" (OuterVolumeSpecName: "kube-api-access-d95bq") pod "6dce6778-383f-4471-b852-16eaed68e455" (UID: "6dce6778-383f-4471-b852-16eaed68e455"). InnerVolumeSpecName "kube-api-access-d95bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.013032 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-xqcwk"] Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.016203 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d95bq\" (UniqueName: \"kubernetes.io/projected/6dce6778-383f-4471-b852-16eaed68e455-kube-api-access-d95bq\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:55 crc kubenswrapper[4810]: W1008 06:47:55.022687 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3b1c5e5_07c2_4c7e_9a71_bd427886615f.slice/crio-0ecd67f7f73d631488f5153082c753ebb5d4e10ce4146f0818533dce00cd5bb7 WatchSource:0}: Error finding container 0ecd67f7f73d631488f5153082c753ebb5d4e10ce4146f0818533dce00cd5bb7: Status 404 returned error can't find the container with id 0ecd67f7f73d631488f5153082c753ebb5d4e10ce4146f0818533dce00cd5bb7 Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.224117 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.225259 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e37514-b509-4b91-b3c5-7a30006a4322" containerName="dnsmasq-dns" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.225303 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e37514-b509-4b91-b3c5-7a30006a4322" containerName="dnsmasq-dns" Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.225336 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22e37514-b509-4b91-b3c5-7a30006a4322" containerName="init" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.225348 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="22e37514-b509-4b91-b3c5-7a30006a4322" containerName="init" Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.225376 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dce6778-383f-4471-b852-16eaed68e455" containerName="mariadb-database-create" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.225388 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dce6778-383f-4471-b852-16eaed68e455" containerName="mariadb-database-create" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.225713 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="22e37514-b509-4b91-b3c5-7a30006a4322" containerName="dnsmasq-dns" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.225755 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dce6778-383f-4471-b852-16eaed68e455" containerName="mariadb-database-create" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.236777 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.239718 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.239940 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.240139 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.240345 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mqsph" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.257180 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.321908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.322014 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.322062 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-lock\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.322090 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfvmk\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-kube-api-access-rfvmk\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.322126 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-cache\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.381092 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfjh7" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.381075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kfjh7" event={"ID":"6dce6778-383f-4471-b852-16eaed68e455","Type":"ContainerDied","Data":"709c082754f29d82d3099868563c0442d68e2646c540edd82f5309992958b1f2"} Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.381144 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="709c082754f29d82d3099868563c0442d68e2646c540edd82f5309992958b1f2" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.383180 4810 generic.go:334] "Generic (PLEG): container finished" podID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerID="7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32" exitCode=0 Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.383211 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" event={"ID":"b3b1c5e5-07c2-4c7e-9a71-bd427886615f","Type":"ContainerDied","Data":"7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32"} Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.383333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" event={"ID":"b3b1c5e5-07c2-4c7e-9a71-bd427886615f","Type":"ContainerStarted","Data":"0ecd67f7f73d631488f5153082c753ebb5d4e10ce4146f0818533dce00cd5bb7"} Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.386659 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.390132 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-d6hjw" event={"ID":"22e37514-b509-4b91-b3c5-7a30006a4322","Type":"ContainerDied","Data":"4617bea7d2120f432a392fbf6dd130c534149fc427d8d5bb0ee0833b2acd145f"} Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.390214 4810 scope.go:117] "RemoveContainer" containerID="61b720b6445fc80efcd5568efd1f845e0db4dbd75a26d415920ab472a986d592" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.423642 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-lock\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.423713 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfvmk\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-kube-api-access-rfvmk\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.423770 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-cache\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.423856 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.423949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.424108 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.424132 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.424180 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift podName:63aa00a9-34a1-45ce-9ada-c767d07eabb8 nodeName:}" failed. No retries permitted until 2025-10-08 06:47:55.924162277 +0000 UTC m=+978.558602007 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift") pod "swift-storage-0" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8") : configmap "swift-ring-files" not found Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.424322 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-lock\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.424401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-cache\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.424581 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.461004 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfvmk\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-kube-api-access-rfvmk\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.466783 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.566370 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-d6hjw"] Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.567640 4810 scope.go:117] "RemoveContainer" containerID="f9650e8f4bf603d068c24d042efbc2d4b597e5c795159be5d3e9daed83ee6a6d" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.575837 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-d6hjw"] Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.658532 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hx4fs" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.731273 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tmw2\" (UniqueName: \"kubernetes.io/projected/508940e7-94fa-4ada-a7b6-0082a5fa6091-kube-api-access-4tmw2\") pod \"508940e7-94fa-4ada-a7b6-0082a5fa6091\" (UID: \"508940e7-94fa-4ada-a7b6-0082a5fa6091\") " Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.740301 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508940e7-94fa-4ada-a7b6-0082a5fa6091-kube-api-access-4tmw2" (OuterVolumeSpecName: "kube-api-access-4tmw2") pod "508940e7-94fa-4ada-a7b6-0082a5fa6091" (UID: "508940e7-94fa-4ada-a7b6-0082a5fa6091"). InnerVolumeSpecName "kube-api-access-4tmw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.741459 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9knv8"] Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.741885 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508940e7-94fa-4ada-a7b6-0082a5fa6091" containerName="mariadb-database-create" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.741978 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="508940e7-94fa-4ada-a7b6-0082a5fa6091" containerName="mariadb-database-create" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.742214 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="508940e7-94fa-4ada-a7b6-0082a5fa6091" containerName="mariadb-database-create" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.742956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.746025 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.746046 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.746547 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.753865 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9knv8"] Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833202 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckm7s\" (UniqueName: \"kubernetes.io/projected/0805f157-1607-41e2-9069-4e07dc59a46b-kube-api-access-ckm7s\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833310 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-scripts\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-combined-ca-bundle\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833378 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0805f157-1607-41e2-9069-4e07dc59a46b-etc-swift\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833453 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-swiftconf\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833479 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-ring-data-devices\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833542 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-dispersionconf\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.833991 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tmw2\" (UniqueName: \"kubernetes.io/projected/508940e7-94fa-4ada-a7b6-0082a5fa6091-kube-api-access-4tmw2\") on node \"crc\" DevicePath \"\"" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935659 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckm7s\" (UniqueName: \"kubernetes.io/projected/0805f157-1607-41e2-9069-4e07dc59a46b-kube-api-access-ckm7s\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935742 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-scripts\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-combined-ca-bundle\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0805f157-1607-41e2-9069-4e07dc59a46b-etc-swift\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935847 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-swiftconf\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935869 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-ring-data-devices\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935887 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-dispersionconf\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.935934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.936129 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.936144 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 06:47:55 crc kubenswrapper[4810]: E1008 06:47:55.936189 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift podName:63aa00a9-34a1-45ce-9ada-c767d07eabb8 nodeName:}" failed. No retries permitted until 2025-10-08 06:47:56.936172613 +0000 UTC m=+979.570612353 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift") pod "swift-storage-0" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8") : configmap "swift-ring-files" not found Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.936342 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0805f157-1607-41e2-9069-4e07dc59a46b-etc-swift\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.936572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-scripts\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.937102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-ring-data-devices\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.940691 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-swiftconf\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.941374 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-dispersionconf\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.943293 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-combined-ca-bundle\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:55 crc kubenswrapper[4810]: I1008 06:47:55.955183 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckm7s\" (UniqueName: \"kubernetes.io/projected/0805f157-1607-41e2-9069-4e07dc59a46b-kube-api-access-ckm7s\") pod \"swift-ring-rebalance-9knv8\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.066310 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.084855 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22e37514-b509-4b91-b3c5-7a30006a4322" path="/var/lib/kubelet/pods/22e37514-b509-4b91-b3c5-7a30006a4322/volumes" Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.405035 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hx4fs" event={"ID":"508940e7-94fa-4ada-a7b6-0082a5fa6091","Type":"ContainerDied","Data":"b434e9828be4313500927eeee24f26abae2d8f725b1f15bbe5dbaa195e3d1fca"} Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.405452 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b434e9828be4313500927eeee24f26abae2d8f725b1f15bbe5dbaa195e3d1fca" Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.405106 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hx4fs" Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.407428 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" event={"ID":"b3b1c5e5-07c2-4c7e-9a71-bd427886615f","Type":"ContainerStarted","Data":"5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348"} Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.407946 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.436454 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" podStartSLOduration=2.436424853 podStartE2EDuration="2.436424853s" podCreationTimestamp="2025-10-08 06:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:47:56.424712358 +0000 UTC m=+979.059152108" watchObservedRunningTime="2025-10-08 06:47:56.436424853 +0000 UTC m=+979.070864603" Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.559316 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9knv8"] Oct 08 06:47:56 crc kubenswrapper[4810]: W1008 06:47:56.566156 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0805f157_1607_41e2_9069_4e07dc59a46b.slice/crio-0c2de8ff46e335c6f9dabf298bfd74d32c1fb623b49c69620ff4ec587ae79d88 WatchSource:0}: Error finding container 0c2de8ff46e335c6f9dabf298bfd74d32c1fb623b49c69620ff4ec587ae79d88: Status 404 returned error can't find the container with id 0c2de8ff46e335c6f9dabf298bfd74d32c1fb623b49c69620ff4ec587ae79d88 Oct 08 06:47:56 crc kubenswrapper[4810]: I1008 06:47:56.959888 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:56 crc kubenswrapper[4810]: E1008 06:47:56.960152 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 06:47:56 crc kubenswrapper[4810]: E1008 06:47:56.960187 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 06:47:56 crc kubenswrapper[4810]: E1008 06:47:56.960265 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift podName:63aa00a9-34a1-45ce-9ada-c767d07eabb8 nodeName:}" failed. No retries permitted until 2025-10-08 06:47:58.960246737 +0000 UTC m=+981.594686477 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift") pod "swift-storage-0" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8") : configmap "swift-ring-files" not found Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.282459 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.427325 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9knv8" event={"ID":"0805f157-1607-41e2-9069-4e07dc59a46b","Type":"ContainerStarted","Data":"0c2de8ff46e335c6f9dabf298bfd74d32c1fb623b49c69620ff4ec587ae79d88"} Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.462316 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-jx9km"] Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.466489 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jx9km" Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.519723 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jx9km"] Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.574574 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxnrz\" (UniqueName: \"kubernetes.io/projected/1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e-kube-api-access-zxnrz\") pod \"glance-db-create-jx9km\" (UID: \"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e\") " pod="openstack/glance-db-create-jx9km" Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.677206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxnrz\" (UniqueName: \"kubernetes.io/projected/1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e-kube-api-access-zxnrz\") pod \"glance-db-create-jx9km\" (UID: \"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e\") " pod="openstack/glance-db-create-jx9km" Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.701068 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxnrz\" (UniqueName: \"kubernetes.io/projected/1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e-kube-api-access-zxnrz\") pod \"glance-db-create-jx9km\" (UID: \"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e\") " pod="openstack/glance-db-create-jx9km" Oct 08 06:47:57 crc kubenswrapper[4810]: I1008 06:47:57.787143 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jx9km" Oct 08 06:47:58 crc kubenswrapper[4810]: I1008 06:47:58.328488 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jx9km"] Oct 08 06:47:58 crc kubenswrapper[4810]: I1008 06:47:58.437803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jx9km" event={"ID":"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e","Type":"ContainerStarted","Data":"40b23b57458ee1fab98e289960d5e61a57b7128fad4b67e1d14e7436c61ccc50"} Oct 08 06:47:59 crc kubenswrapper[4810]: I1008 06:47:59.006741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:47:59 crc kubenswrapper[4810]: E1008 06:47:59.007028 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 06:47:59 crc kubenswrapper[4810]: E1008 06:47:59.007472 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 06:47:59 crc kubenswrapper[4810]: E1008 06:47:59.007549 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift podName:63aa00a9-34a1-45ce-9ada-c767d07eabb8 nodeName:}" failed. No retries permitted until 2025-10-08 06:48:03.00752232 +0000 UTC m=+985.641962060 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift") pod "swift-storage-0" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8") : configmap "swift-ring-files" not found Oct 08 06:47:59 crc kubenswrapper[4810]: I1008 06:47:59.449641 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e" containerID="467ac7e064869303b6ece8fd50d804b1d6f99a7fadaee0b3b8c2839cf2711511" exitCode=0 Oct 08 06:47:59 crc kubenswrapper[4810]: I1008 06:47:59.449703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jx9km" event={"ID":"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e","Type":"ContainerDied","Data":"467ac7e064869303b6ece8fd50d804b1d6f99a7fadaee0b3b8c2839cf2711511"} Oct 08 06:48:01 crc kubenswrapper[4810]: I1008 06:48:01.322579 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jx9km" Oct 08 06:48:01 crc kubenswrapper[4810]: I1008 06:48:01.465414 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxnrz\" (UniqueName: \"kubernetes.io/projected/1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e-kube-api-access-zxnrz\") pod \"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e\" (UID: \"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e\") " Oct 08 06:48:01 crc kubenswrapper[4810]: I1008 06:48:01.476076 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jx9km" event={"ID":"1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e","Type":"ContainerDied","Data":"40b23b57458ee1fab98e289960d5e61a57b7128fad4b67e1d14e7436c61ccc50"} Oct 08 06:48:01 crc kubenswrapper[4810]: I1008 06:48:01.476611 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40b23b57458ee1fab98e289960d5e61a57b7128fad4b67e1d14e7436c61ccc50" Oct 08 06:48:01 crc kubenswrapper[4810]: I1008 06:48:01.476195 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jx9km" Oct 08 06:48:01 crc kubenswrapper[4810]: I1008 06:48:01.479377 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e-kube-api-access-zxnrz" (OuterVolumeSpecName: "kube-api-access-zxnrz") pod "1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e" (UID: "1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e"). InnerVolumeSpecName "kube-api-access-zxnrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:01 crc kubenswrapper[4810]: I1008 06:48:01.568678 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxnrz\" (UniqueName: \"kubernetes.io/projected/1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e-kube-api-access-zxnrz\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.000740 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-eac1-account-create-bqdsm"] Oct 08 06:48:02 crc kubenswrapper[4810]: E1008 06:48:02.001210 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e" containerName="mariadb-database-create" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.001232 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e" containerName="mariadb-database-create" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.001482 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e" containerName="mariadb-database-create" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.005263 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eac1-account-create-bqdsm" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.007690 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.013613 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eac1-account-create-bqdsm"] Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.079859 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfhcd\" (UniqueName: \"kubernetes.io/projected/e1f667ff-3e8b-48b9-be0d-5deb9483050a-kube-api-access-qfhcd\") pod \"keystone-eac1-account-create-bqdsm\" (UID: \"e1f667ff-3e8b-48b9-be0d-5deb9483050a\") " pod="openstack/keystone-eac1-account-create-bqdsm" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.181876 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfhcd\" (UniqueName: \"kubernetes.io/projected/e1f667ff-3e8b-48b9-be0d-5deb9483050a-kube-api-access-qfhcd\") pod \"keystone-eac1-account-create-bqdsm\" (UID: \"e1f667ff-3e8b-48b9-be0d-5deb9483050a\") " pod="openstack/keystone-eac1-account-create-bqdsm" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.201959 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfhcd\" (UniqueName: \"kubernetes.io/projected/e1f667ff-3e8b-48b9-be0d-5deb9483050a-kube-api-access-qfhcd\") pod \"keystone-eac1-account-create-bqdsm\" (UID: \"e1f667ff-3e8b-48b9-be0d-5deb9483050a\") " pod="openstack/keystone-eac1-account-create-bqdsm" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.322120 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c68e-account-create-rz9v2"] Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.323233 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c68e-account-create-rz9v2" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.332096 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c68e-account-create-rz9v2"] Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.332541 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.332895 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eac1-account-create-bqdsm" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.388241 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.389255 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mrxq\" (UniqueName: \"kubernetes.io/projected/d12c373e-80ef-4de3-bb28-1044cfa4b8d6-kube-api-access-5mrxq\") pod \"placement-c68e-account-create-rz9v2\" (UID: \"d12c373e-80ef-4de3-bb28-1044cfa4b8d6\") " pod="openstack/placement-c68e-account-create-rz9v2" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.491804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mrxq\" (UniqueName: \"kubernetes.io/projected/d12c373e-80ef-4de3-bb28-1044cfa4b8d6-kube-api-access-5mrxq\") pod \"placement-c68e-account-create-rz9v2\" (UID: \"d12c373e-80ef-4de3-bb28-1044cfa4b8d6\") " pod="openstack/placement-c68e-account-create-rz9v2" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.492891 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9knv8" event={"ID":"0805f157-1607-41e2-9069-4e07dc59a46b","Type":"ContainerStarted","Data":"304adddb1efc0687866b5c780e8dddddb285afdc0117a61a17f3af47c6303e9f"} Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.527717 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9knv8" podStartSLOduration=2.890058345 podStartE2EDuration="7.527687689s" podCreationTimestamp="2025-10-08 06:47:55 +0000 UTC" firstStartedPulling="2025-10-08 06:47:56.567803808 +0000 UTC m=+979.202243548" lastFinishedPulling="2025-10-08 06:48:01.205433142 +0000 UTC m=+983.839872892" observedRunningTime="2025-10-08 06:48:02.525664554 +0000 UTC m=+985.160104294" watchObservedRunningTime="2025-10-08 06:48:02.527687689 +0000 UTC m=+985.162127429" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.528954 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mrxq\" (UniqueName: \"kubernetes.io/projected/d12c373e-80ef-4de3-bb28-1044cfa4b8d6-kube-api-access-5mrxq\") pod \"placement-c68e-account-create-rz9v2\" (UID: \"d12c373e-80ef-4de3-bb28-1044cfa4b8d6\") " pod="openstack/placement-c68e-account-create-rz9v2" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.647728 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c68e-account-create-rz9v2" Oct 08 06:48:02 crc kubenswrapper[4810]: I1008 06:48:02.931147 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-eac1-account-create-bqdsm"] Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.106267 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:48:03 crc kubenswrapper[4810]: E1008 06:48:03.106517 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 06:48:03 crc kubenswrapper[4810]: E1008 06:48:03.106933 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 06:48:03 crc kubenswrapper[4810]: E1008 06:48:03.107009 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift podName:63aa00a9-34a1-45ce-9ada-c767d07eabb8 nodeName:}" failed. No retries permitted until 2025-10-08 06:48:11.106990432 +0000 UTC m=+993.741430172 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift") pod "swift-storage-0" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8") : configmap "swift-ring-files" not found Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.189591 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c68e-account-create-rz9v2"] Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.506550 4810 generic.go:334] "Generic (PLEG): container finished" podID="e1f667ff-3e8b-48b9-be0d-5deb9483050a" containerID="95cc241da1c907b8dff8d4520a8e079f4946bdfb0285b71777b56e61b76c747e" exitCode=0 Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.506661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eac1-account-create-bqdsm" event={"ID":"e1f667ff-3e8b-48b9-be0d-5deb9483050a","Type":"ContainerDied","Data":"95cc241da1c907b8dff8d4520a8e079f4946bdfb0285b71777b56e61b76c747e"} Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.506891 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eac1-account-create-bqdsm" event={"ID":"e1f667ff-3e8b-48b9-be0d-5deb9483050a","Type":"ContainerStarted","Data":"9c42664ecdea7ed4bffa4f4f797941e166b7cd46b506289f7c62b0abf068d5eb"} Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.508926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c68e-account-create-rz9v2" event={"ID":"d12c373e-80ef-4de3-bb28-1044cfa4b8d6","Type":"ContainerStarted","Data":"9c36ae631cf8df1c181b11732cc17ae7be213aa863fab14cd460920c2dd2e53d"} Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.509084 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c68e-account-create-rz9v2" event={"ID":"d12c373e-80ef-4de3-bb28-1044cfa4b8d6","Type":"ContainerStarted","Data":"93e7b783fa93f8a53bf7f54f0afcbfae9d56e5193895fae80be72f88f6ccc1ef"} Oct 08 06:48:03 crc kubenswrapper[4810]: I1008 06:48:03.552701 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-c68e-account-create-rz9v2" podStartSLOduration=1.552681578 podStartE2EDuration="1.552681578s" podCreationTimestamp="2025-10-08 06:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:03.551675081 +0000 UTC m=+986.186114821" watchObservedRunningTime="2025-10-08 06:48:03.552681578 +0000 UTC m=+986.187121318" Oct 08 06:48:04 crc kubenswrapper[4810]: I1008 06:48:04.433284 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:48:04 crc kubenswrapper[4810]: I1008 06:48:04.516339 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-nmfc4"] Oct 08 06:48:04 crc kubenswrapper[4810]: I1008 06:48:04.516669 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" podUID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerName="dnsmasq-dns" containerID="cri-o://665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea" gracePeriod=10 Oct 08 06:48:04 crc kubenswrapper[4810]: I1008 06:48:04.576676 4810 generic.go:334] "Generic (PLEG): container finished" podID="d12c373e-80ef-4de3-bb28-1044cfa4b8d6" containerID="9c36ae631cf8df1c181b11732cc17ae7be213aa863fab14cd460920c2dd2e53d" exitCode=0 Oct 08 06:48:04 crc kubenswrapper[4810]: I1008 06:48:04.576785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c68e-account-create-rz9v2" event={"ID":"d12c373e-80ef-4de3-bb28-1044cfa4b8d6","Type":"ContainerDied","Data":"9c36ae631cf8df1c181b11732cc17ae7be213aa863fab14cd460920c2dd2e53d"} Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.079922 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eac1-account-create-bqdsm" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.086006 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.171052 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-dns-svc\") pod \"df9f8bfe-4f8d-4f95-910c-0065aa227782\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.171123 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-nb\") pod \"df9f8bfe-4f8d-4f95-910c-0065aa227782\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.171214 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-config\") pod \"df9f8bfe-4f8d-4f95-910c-0065aa227782\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.171325 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bprw5\" (UniqueName: \"kubernetes.io/projected/df9f8bfe-4f8d-4f95-910c-0065aa227782-kube-api-access-bprw5\") pod \"df9f8bfe-4f8d-4f95-910c-0065aa227782\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.171393 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfhcd\" (UniqueName: \"kubernetes.io/projected/e1f667ff-3e8b-48b9-be0d-5deb9483050a-kube-api-access-qfhcd\") pod \"e1f667ff-3e8b-48b9-be0d-5deb9483050a\" (UID: \"e1f667ff-3e8b-48b9-be0d-5deb9483050a\") " Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.171506 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-sb\") pod \"df9f8bfe-4f8d-4f95-910c-0065aa227782\" (UID: \"df9f8bfe-4f8d-4f95-910c-0065aa227782\") " Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.184487 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df9f8bfe-4f8d-4f95-910c-0065aa227782-kube-api-access-bprw5" (OuterVolumeSpecName: "kube-api-access-bprw5") pod "df9f8bfe-4f8d-4f95-910c-0065aa227782" (UID: "df9f8bfe-4f8d-4f95-910c-0065aa227782"). InnerVolumeSpecName "kube-api-access-bprw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.193231 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f667ff-3e8b-48b9-be0d-5deb9483050a-kube-api-access-qfhcd" (OuterVolumeSpecName: "kube-api-access-qfhcd") pod "e1f667ff-3e8b-48b9-be0d-5deb9483050a" (UID: "e1f667ff-3e8b-48b9-be0d-5deb9483050a"). InnerVolumeSpecName "kube-api-access-qfhcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.227022 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-config" (OuterVolumeSpecName: "config") pod "df9f8bfe-4f8d-4f95-910c-0065aa227782" (UID: "df9f8bfe-4f8d-4f95-910c-0065aa227782"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.234471 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df9f8bfe-4f8d-4f95-910c-0065aa227782" (UID: "df9f8bfe-4f8d-4f95-910c-0065aa227782"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.237127 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df9f8bfe-4f8d-4f95-910c-0065aa227782" (UID: "df9f8bfe-4f8d-4f95-910c-0065aa227782"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.244359 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df9f8bfe-4f8d-4f95-910c-0065aa227782" (UID: "df9f8bfe-4f8d-4f95-910c-0065aa227782"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.275587 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.275929 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.276086 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.276188 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bprw5\" (UniqueName: \"kubernetes.io/projected/df9f8bfe-4f8d-4f95-910c-0065aa227782-kube-api-access-bprw5\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.276289 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfhcd\" (UniqueName: \"kubernetes.io/projected/e1f667ff-3e8b-48b9-be0d-5deb9483050a-kube-api-access-qfhcd\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.276383 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df9f8bfe-4f8d-4f95-910c-0065aa227782-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.592285 4810 generic.go:334] "Generic (PLEG): container finished" podID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerID="665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea" exitCode=0 Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.592383 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" event={"ID":"df9f8bfe-4f8d-4f95-910c-0065aa227782","Type":"ContainerDied","Data":"665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea"} Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.592430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" event={"ID":"df9f8bfe-4f8d-4f95-910c-0065aa227782","Type":"ContainerDied","Data":"08ef1d016c592ff78e29f02e93e848b41352d6fc78b6b60d1175cf73859a704d"} Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.592422 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-nmfc4" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.592459 4810 scope.go:117] "RemoveContainer" containerID="665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.596472 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-eac1-account-create-bqdsm" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.597173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-eac1-account-create-bqdsm" event={"ID":"e1f667ff-3e8b-48b9-be0d-5deb9483050a","Type":"ContainerDied","Data":"9c42664ecdea7ed4bffa4f4f797941e166b7cd46b506289f7c62b0abf068d5eb"} Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.597251 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c42664ecdea7ed4bffa4f4f797941e166b7cd46b506289f7c62b0abf068d5eb" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.636703 4810 scope.go:117] "RemoveContainer" containerID="9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.663954 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-nmfc4"] Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.674282 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-nmfc4"] Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.746800 4810 scope.go:117] "RemoveContainer" containerID="665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea" Oct 08 06:48:05 crc kubenswrapper[4810]: E1008 06:48:05.747624 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea\": container with ID starting with 665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea not found: ID does not exist" containerID="665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.747691 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea"} err="failed to get container status \"665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea\": rpc error: code = NotFound desc = could not find container \"665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea\": container with ID starting with 665ca3e656ea7c0d651b3a80930ff7a491e45b982f2e679f90b90aaae9ee6eea not found: ID does not exist" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.747765 4810 scope.go:117] "RemoveContainer" containerID="9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b" Oct 08 06:48:05 crc kubenswrapper[4810]: E1008 06:48:05.748183 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b\": container with ID starting with 9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b not found: ID does not exist" containerID="9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b" Oct 08 06:48:05 crc kubenswrapper[4810]: I1008 06:48:05.748205 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b"} err="failed to get container status \"9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b\": rpc error: code = NotFound desc = could not find container \"9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b\": container with ID starting with 9c856d32375e1c025015b88a3a3509d475a9fb93e428f6a7c0c46be05348315b not found: ID does not exist" Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.027479 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c68e-account-create-rz9v2" Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.087085 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df9f8bfe-4f8d-4f95-910c-0065aa227782" path="/var/lib/kubelet/pods/df9f8bfe-4f8d-4f95-910c-0065aa227782/volumes" Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.090894 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mrxq\" (UniqueName: \"kubernetes.io/projected/d12c373e-80ef-4de3-bb28-1044cfa4b8d6-kube-api-access-5mrxq\") pod \"d12c373e-80ef-4de3-bb28-1044cfa4b8d6\" (UID: \"d12c373e-80ef-4de3-bb28-1044cfa4b8d6\") " Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.096564 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d12c373e-80ef-4de3-bb28-1044cfa4b8d6-kube-api-access-5mrxq" (OuterVolumeSpecName: "kube-api-access-5mrxq") pod "d12c373e-80ef-4de3-bb28-1044cfa4b8d6" (UID: "d12c373e-80ef-4de3-bb28-1044cfa4b8d6"). InnerVolumeSpecName "kube-api-access-5mrxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.197195 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mrxq\" (UniqueName: \"kubernetes.io/projected/d12c373e-80ef-4de3-bb28-1044cfa4b8d6-kube-api-access-5mrxq\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.613084 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c68e-account-create-rz9v2" Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.613070 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c68e-account-create-rz9v2" event={"ID":"d12c373e-80ef-4de3-bb28-1044cfa4b8d6","Type":"ContainerDied","Data":"93e7b783fa93f8a53bf7f54f0afcbfae9d56e5193895fae80be72f88f6ccc1ef"} Oct 08 06:48:06 crc kubenswrapper[4810]: I1008 06:48:06.614003 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93e7b783fa93f8a53bf7f54f0afcbfae9d56e5193895fae80be72f88f6ccc1ef" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.536220 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f3d6-account-create-n49kn"] Oct 08 06:48:07 crc kubenswrapper[4810]: E1008 06:48:07.537142 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerName="init" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.537170 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerName="init" Oct 08 06:48:07 crc kubenswrapper[4810]: E1008 06:48:07.537193 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerName="dnsmasq-dns" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.537206 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerName="dnsmasq-dns" Oct 08 06:48:07 crc kubenswrapper[4810]: E1008 06:48:07.537249 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f667ff-3e8b-48b9-be0d-5deb9483050a" containerName="mariadb-account-create" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.537259 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f667ff-3e8b-48b9-be0d-5deb9483050a" containerName="mariadb-account-create" Oct 08 06:48:07 crc kubenswrapper[4810]: E1008 06:48:07.537272 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d12c373e-80ef-4de3-bb28-1044cfa4b8d6" containerName="mariadb-account-create" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.537280 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d12c373e-80ef-4de3-bb28-1044cfa4b8d6" containerName="mariadb-account-create" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.537500 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1f667ff-3e8b-48b9-be0d-5deb9483050a" containerName="mariadb-account-create" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.537515 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d12c373e-80ef-4de3-bb28-1044cfa4b8d6" containerName="mariadb-account-create" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.537536 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="df9f8bfe-4f8d-4f95-910c-0065aa227782" containerName="dnsmasq-dns" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.538346 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f3d6-account-create-n49kn" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.541303 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.552413 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f3d6-account-create-n49kn"] Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.626363 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkkm6\" (UniqueName: \"kubernetes.io/projected/8076563f-a890-4f24-8b45-c4141d27fa73-kube-api-access-kkkm6\") pod \"glance-f3d6-account-create-n49kn\" (UID: \"8076563f-a890-4f24-8b45-c4141d27fa73\") " pod="openstack/glance-f3d6-account-create-n49kn" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.628281 4810 generic.go:334] "Generic (PLEG): container finished" podID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerID="a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1" exitCode=0 Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.628352 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2a44c96c-b96c-409c-9c9e-c049d9fe68b5","Type":"ContainerDied","Data":"a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1"} Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.639811 4810 generic.go:334] "Generic (PLEG): container finished" podID="4deb0612-547f-4067-b95b-5794663d21aa" containerID="c3755a82f85d0942937435d7ed4103b19749825c63d0c7188863bb653199761b" exitCode=0 Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.639887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4deb0612-547f-4067-b95b-5794663d21aa","Type":"ContainerDied","Data":"c3755a82f85d0942937435d7ed4103b19749825c63d0c7188863bb653199761b"} Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.731191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkkm6\" (UniqueName: \"kubernetes.io/projected/8076563f-a890-4f24-8b45-c4141d27fa73-kube-api-access-kkkm6\") pod \"glance-f3d6-account-create-n49kn\" (UID: \"8076563f-a890-4f24-8b45-c4141d27fa73\") " pod="openstack/glance-f3d6-account-create-n49kn" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.769104 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkkm6\" (UniqueName: \"kubernetes.io/projected/8076563f-a890-4f24-8b45-c4141d27fa73-kube-api-access-kkkm6\") pod \"glance-f3d6-account-create-n49kn\" (UID: \"8076563f-a890-4f24-8b45-c4141d27fa73\") " pod="openstack/glance-f3d6-account-create-n49kn" Oct 08 06:48:07 crc kubenswrapper[4810]: I1008 06:48:07.861127 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f3d6-account-create-n49kn" Oct 08 06:48:08 crc kubenswrapper[4810]: I1008 06:48:08.358434 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f3d6-account-create-n49kn"] Oct 08 06:48:08 crc kubenswrapper[4810]: I1008 06:48:08.652256 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f3d6-account-create-n49kn" event={"ID":"8076563f-a890-4f24-8b45-c4141d27fa73","Type":"ContainerStarted","Data":"50273996e670a01641009752cebdad62b933c293da59c8a34e358694a1d1b4e7"} Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.671314 4810 generic.go:334] "Generic (PLEG): container finished" podID="8076563f-a890-4f24-8b45-c4141d27fa73" containerID="7456a6a32595ebd5c9681d2ead99703593de54cb42b0b538d1d12ebd1daeb457" exitCode=0 Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.671750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f3d6-account-create-n49kn" event={"ID":"8076563f-a890-4f24-8b45-c4141d27fa73","Type":"ContainerDied","Data":"7456a6a32595ebd5c9681d2ead99703593de54cb42b0b538d1d12ebd1daeb457"} Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.673721 4810 generic.go:334] "Generic (PLEG): container finished" podID="0805f157-1607-41e2-9069-4e07dc59a46b" containerID="304adddb1efc0687866b5c780e8dddddb285afdc0117a61a17f3af47c6303e9f" exitCode=0 Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.673776 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9knv8" event={"ID":"0805f157-1607-41e2-9069-4e07dc59a46b","Type":"ContainerDied","Data":"304adddb1efc0687866b5c780e8dddddb285afdc0117a61a17f3af47c6303e9f"} Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.677669 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2a44c96c-b96c-409c-9c9e-c049d9fe68b5","Type":"ContainerStarted","Data":"5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98"} Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.677896 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.680067 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4deb0612-547f-4067-b95b-5794663d21aa","Type":"ContainerStarted","Data":"2d507eb2a2ecc2839157b138e7210f5bbddda79641955e41f82171d217de786c"} Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.680272 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.719781 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.24096585 podStartE2EDuration="1m2.719758388s" podCreationTimestamp="2025-10-08 06:47:07 +0000 UTC" firstStartedPulling="2025-10-08 06:47:09.795413625 +0000 UTC m=+932.429853365" lastFinishedPulling="2025-10-08 06:47:33.274206163 +0000 UTC m=+955.908645903" observedRunningTime="2025-10-08 06:48:09.712072224 +0000 UTC m=+992.346511984" watchObservedRunningTime="2025-10-08 06:48:09.719758388 +0000 UTC m=+992.354198128" Oct 08 06:48:09 crc kubenswrapper[4810]: I1008 06:48:09.766270 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.908016947 podStartE2EDuration="1m2.766251027s" podCreationTimestamp="2025-10-08 06:47:07 +0000 UTC" firstStartedPulling="2025-10-08 06:47:09.227686652 +0000 UTC m=+931.862126392" lastFinishedPulling="2025-10-08 06:47:33.085920732 +0000 UTC m=+955.720360472" observedRunningTime="2025-10-08 06:48:09.743097725 +0000 UTC m=+992.377537485" watchObservedRunningTime="2025-10-08 06:48:09.766251027 +0000 UTC m=+992.400690757" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.124625 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.132597 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f3d6-account-create-n49kn" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.199548 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.210345 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"swift-storage-0\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " pod="openstack/swift-storage-0" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.301803 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0805f157-1607-41e2-9069-4e07dc59a46b-etc-swift\") pod \"0805f157-1607-41e2-9069-4e07dc59a46b\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.301940 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-scripts\") pod \"0805f157-1607-41e2-9069-4e07dc59a46b\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.302051 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-swiftconf\") pod \"0805f157-1607-41e2-9069-4e07dc59a46b\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.302125 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckm7s\" (UniqueName: \"kubernetes.io/projected/0805f157-1607-41e2-9069-4e07dc59a46b-kube-api-access-ckm7s\") pod \"0805f157-1607-41e2-9069-4e07dc59a46b\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.302279 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-combined-ca-bundle\") pod \"0805f157-1607-41e2-9069-4e07dc59a46b\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.302896 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0805f157-1607-41e2-9069-4e07dc59a46b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0805f157-1607-41e2-9069-4e07dc59a46b" (UID: "0805f157-1607-41e2-9069-4e07dc59a46b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.303105 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkkm6\" (UniqueName: \"kubernetes.io/projected/8076563f-a890-4f24-8b45-c4141d27fa73-kube-api-access-kkkm6\") pod \"8076563f-a890-4f24-8b45-c4141d27fa73\" (UID: \"8076563f-a890-4f24-8b45-c4141d27fa73\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.303234 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-dispersionconf\") pod \"0805f157-1607-41e2-9069-4e07dc59a46b\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.303287 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-ring-data-devices\") pod \"0805f157-1607-41e2-9069-4e07dc59a46b\" (UID: \"0805f157-1607-41e2-9069-4e07dc59a46b\") " Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.304246 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0805f157-1607-41e2-9069-4e07dc59a46b-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.304582 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0805f157-1607-41e2-9069-4e07dc59a46b" (UID: "0805f157-1607-41e2-9069-4e07dc59a46b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.315489 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0805f157-1607-41e2-9069-4e07dc59a46b-kube-api-access-ckm7s" (OuterVolumeSpecName: "kube-api-access-ckm7s") pod "0805f157-1607-41e2-9069-4e07dc59a46b" (UID: "0805f157-1607-41e2-9069-4e07dc59a46b"). InnerVolumeSpecName "kube-api-access-ckm7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.317109 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8076563f-a890-4f24-8b45-c4141d27fa73-kube-api-access-kkkm6" (OuterVolumeSpecName: "kube-api-access-kkkm6") pod "8076563f-a890-4f24-8b45-c4141d27fa73" (UID: "8076563f-a890-4f24-8b45-c4141d27fa73"). InnerVolumeSpecName "kube-api-access-kkkm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.324833 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0805f157-1607-41e2-9069-4e07dc59a46b" (UID: "0805f157-1607-41e2-9069-4e07dc59a46b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.329093 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-scripts" (OuterVolumeSpecName: "scripts") pod "0805f157-1607-41e2-9069-4e07dc59a46b" (UID: "0805f157-1607-41e2-9069-4e07dc59a46b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.331438 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0805f157-1607-41e2-9069-4e07dc59a46b" (UID: "0805f157-1607-41e2-9069-4e07dc59a46b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.348073 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0805f157-1607-41e2-9069-4e07dc59a46b" (UID: "0805f157-1607-41e2-9069-4e07dc59a46b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.413188 4810 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.413253 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckm7s\" (UniqueName: \"kubernetes.io/projected/0805f157-1607-41e2-9069-4e07dc59a46b-kube-api-access-ckm7s\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.413269 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.413289 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkkm6\" (UniqueName: \"kubernetes.io/projected/8076563f-a890-4f24-8b45-c4141d27fa73-kube-api-access-kkkm6\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.413307 4810 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0805f157-1607-41e2-9069-4e07dc59a46b-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.413317 4810 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.413326 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0805f157-1607-41e2-9069-4e07dc59a46b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.462500 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.702990 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f3d6-account-create-n49kn" event={"ID":"8076563f-a890-4f24-8b45-c4141d27fa73","Type":"ContainerDied","Data":"50273996e670a01641009752cebdad62b933c293da59c8a34e358694a1d1b4e7"} Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.703038 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50273996e670a01641009752cebdad62b933c293da59c8a34e358694a1d1b4e7" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.703104 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f3d6-account-create-n49kn" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.718462 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9knv8" event={"ID":"0805f157-1607-41e2-9069-4e07dc59a46b","Type":"ContainerDied","Data":"0c2de8ff46e335c6f9dabf298bfd74d32c1fb623b49c69620ff4ec587ae79d88"} Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.718515 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c2de8ff46e335c6f9dabf298bfd74d32c1fb623b49c69620ff4ec587ae79d88" Oct 08 06:48:11 crc kubenswrapper[4810]: I1008 06:48:11.718623 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9knv8" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.087471 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.286537 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-qtj9x" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" probeResult="failure" output=< Oct 08 06:48:12 crc kubenswrapper[4810]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 06:48:12 crc kubenswrapper[4810]: > Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.317196 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.334628 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.580612 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-qtj9x-config-tk2gx"] Oct 08 06:48:12 crc kubenswrapper[4810]: E1008 06:48:12.581152 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0805f157-1607-41e2-9069-4e07dc59a46b" containerName="swift-ring-rebalance" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.581172 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0805f157-1607-41e2-9069-4e07dc59a46b" containerName="swift-ring-rebalance" Oct 08 06:48:12 crc kubenswrapper[4810]: E1008 06:48:12.581183 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8076563f-a890-4f24-8b45-c4141d27fa73" containerName="mariadb-account-create" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.581190 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8076563f-a890-4f24-8b45-c4141d27fa73" containerName="mariadb-account-create" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.581408 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8076563f-a890-4f24-8b45-c4141d27fa73" containerName="mariadb-account-create" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.581439 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0805f157-1607-41e2-9069-4e07dc59a46b" containerName="swift-ring-rebalance" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.582144 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.595443 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.595369 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qtj9x-config-tk2gx"] Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.704292 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-lcpxn"] Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.705749 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.709473 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.710038 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6dmpd" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.733377 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-lcpxn"] Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.738227 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqnmc\" (UniqueName: \"kubernetes.io/projected/9318e8c4-6194-4ae9-a7d1-1865a5be9583-kube-api-access-jqnmc\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.738365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.738392 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-scripts\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.738420 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-log-ovn\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.738440 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run-ovn\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.738464 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-additional-scripts\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.739278 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"cfbfca7b2ba7b12331603d4ad82d621e062bc8d874c683ffabae0d62b492524e"} Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-log-ovn\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840390 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run-ovn\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840423 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-additional-scripts\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840456 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-db-sync-config-data\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqnmc\" (UniqueName: \"kubernetes.io/projected/9318e8c4-6194-4ae9-a7d1-1865a5be9583-kube-api-access-jqnmc\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840554 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-combined-ca-bundle\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840839 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run-ovn\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.840918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-log-ovn\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.841263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-config-data\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.841417 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9zwd\" (UniqueName: \"kubernetes.io/projected/c40f9ad6-fed6-4759-8aa3-134912fd46bc-kube-api-access-n9zwd\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.841477 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.841507 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-additional-scripts\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.841522 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-scripts\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.841700 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.843856 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-scripts\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.863309 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqnmc\" (UniqueName: \"kubernetes.io/projected/9318e8c4-6194-4ae9-a7d1-1865a5be9583-kube-api-access-jqnmc\") pod \"ovn-controller-qtj9x-config-tk2gx\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.906833 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.944147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9zwd\" (UniqueName: \"kubernetes.io/projected/c40f9ad6-fed6-4759-8aa3-134912fd46bc-kube-api-access-n9zwd\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.944294 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-db-sync-config-data\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.944397 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-combined-ca-bundle\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.944503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-config-data\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.949999 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-combined-ca-bundle\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.950680 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-db-sync-config-data\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.951842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-config-data\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:12 crc kubenswrapper[4810]: I1008 06:48:12.966069 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9zwd\" (UniqueName: \"kubernetes.io/projected/c40f9ad6-fed6-4759-8aa3-134912fd46bc-kube-api-access-n9zwd\") pod \"glance-db-sync-lcpxn\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:13 crc kubenswrapper[4810]: I1008 06:48:13.032037 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:14 crc kubenswrapper[4810]: I1008 06:48:13.502262 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qtj9x-config-tk2gx"] Oct 08 06:48:14 crc kubenswrapper[4810]: I1008 06:48:13.702642 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-lcpxn"] Oct 08 06:48:14 crc kubenswrapper[4810]: I1008 06:48:14.767376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lcpxn" event={"ID":"c40f9ad6-fed6-4759-8aa3-134912fd46bc","Type":"ContainerStarted","Data":"ebc0e69e8e304e5e88de2824441e3f694967b5bc722a024b0dde1d9ec6097c13"} Oct 08 06:48:14 crc kubenswrapper[4810]: I1008 06:48:14.770324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x-config-tk2gx" event={"ID":"9318e8c4-6194-4ae9-a7d1-1865a5be9583","Type":"ContainerStarted","Data":"8c68488d212e0b001c6dcc8db4093e1f17355124f20eca3138f449ba497246c3"} Oct 08 06:48:15 crc kubenswrapper[4810]: I1008 06:48:15.783623 4810 generic.go:334] "Generic (PLEG): container finished" podID="9318e8c4-6194-4ae9-a7d1-1865a5be9583" containerID="3a567dca81c9cd9b5bcaec9dfaa1354ae1f435226e25175f097dce75fffa0d2e" exitCode=0 Oct 08 06:48:15 crc kubenswrapper[4810]: I1008 06:48:15.783729 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x-config-tk2gx" event={"ID":"9318e8c4-6194-4ae9-a7d1-1865a5be9583","Type":"ContainerDied","Data":"3a567dca81c9cd9b5bcaec9dfaa1354ae1f435226e25175f097dce75fffa0d2e"} Oct 08 06:48:15 crc kubenswrapper[4810]: I1008 06:48:15.789092 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"f2a45c96c2a4d65f22eb75f3b31681ceb8d4be034c534c71d36192573c688352"} Oct 08 06:48:15 crc kubenswrapper[4810]: I1008 06:48:15.789128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"cc83732eb2aa89586c96fc2f8b3a8d349fb000018c5d29cd6d05bd9035d6901c"} Oct 08 06:48:15 crc kubenswrapper[4810]: I1008 06:48:15.789141 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"f97fcf1457257c718d7bc50e69bb8e879b0bcdda1152aafd198dc4eacc3764c1"} Oct 08 06:48:15 crc kubenswrapper[4810]: I1008 06:48:15.789152 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"39e49e136cc29f048dd01f7cdb829f4235eb528d5956a0e907b4abc94d91b6bd"} Oct 08 06:48:16 crc kubenswrapper[4810]: I1008 06:48:16.803007 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"ce160cdae1ce3f85d7b348919fd8ec1e63ee09bafdd24ff6e4c23db7ad5a2ff1"} Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.212820 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.238385 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-log-ovn\") pod \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.238475 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-scripts\") pod \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.238504 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run\") pod \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.238528 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqnmc\" (UniqueName: \"kubernetes.io/projected/9318e8c4-6194-4ae9-a7d1-1865a5be9583-kube-api-access-jqnmc\") pod \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.238693 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run-ovn\") pod \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.238735 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-additional-scripts\") pod \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\" (UID: \"9318e8c4-6194-4ae9-a7d1-1865a5be9583\") " Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.239534 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9318e8c4-6194-4ae9-a7d1-1865a5be9583" (UID: "9318e8c4-6194-4ae9-a7d1-1865a5be9583"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.240447 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9318e8c4-6194-4ae9-a7d1-1865a5be9583" (UID: "9318e8c4-6194-4ae9-a7d1-1865a5be9583"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.240519 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run" (OuterVolumeSpecName: "var-run") pod "9318e8c4-6194-4ae9-a7d1-1865a5be9583" (UID: "9318e8c4-6194-4ae9-a7d1-1865a5be9583"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.240641 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-scripts" (OuterVolumeSpecName: "scripts") pod "9318e8c4-6194-4ae9-a7d1-1865a5be9583" (UID: "9318e8c4-6194-4ae9-a7d1-1865a5be9583"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.251307 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9318e8c4-6194-4ae9-a7d1-1865a5be9583" (UID: "9318e8c4-6194-4ae9-a7d1-1865a5be9583"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.268679 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9318e8c4-6194-4ae9-a7d1-1865a5be9583-kube-api-access-jqnmc" (OuterVolumeSpecName: "kube-api-access-jqnmc") pod "9318e8c4-6194-4ae9-a7d1-1865a5be9583" (UID: "9318e8c4-6194-4ae9-a7d1-1865a5be9583"). InnerVolumeSpecName "kube-api-access-jqnmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.340647 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.340680 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.340713 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.340723 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9318e8c4-6194-4ae9-a7d1-1865a5be9583-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.340732 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9318e8c4-6194-4ae9-a7d1-1865a5be9583-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.340740 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqnmc\" (UniqueName: \"kubernetes.io/projected/9318e8c4-6194-4ae9-a7d1-1865a5be9583-kube-api-access-jqnmc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.347329 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-qtj9x" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.819287 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"b1c09d7235386cf593d3456b3d746d6d1cf962a7a681c42f5eb436bbfa5c7423"} Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.819677 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"aeceb614fb9899cfafb9c3fa078fce2b620e618a9756ff3d62a4a160e354a0b2"} Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.819691 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"8c66da6cd63ffdac911d7e4cb28fc214ac4f2ca0696b970109e6193d96655d61"} Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.821447 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x-config-tk2gx" event={"ID":"9318e8c4-6194-4ae9-a7d1-1865a5be9583","Type":"ContainerDied","Data":"8c68488d212e0b001c6dcc8db4093e1f17355124f20eca3138f449ba497246c3"} Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.821503 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c68488d212e0b001c6dcc8db4093e1f17355124f20eca3138f449ba497246c3" Oct 08 06:48:17 crc kubenswrapper[4810]: I1008 06:48:17.821530 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-tk2gx" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.340526 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-qtj9x-config-tk2gx"] Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.346997 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-qtj9x-config-tk2gx"] Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.453612 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-qtj9x-config-fwvq8"] Oct 08 06:48:18 crc kubenswrapper[4810]: E1008 06:48:18.454168 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9318e8c4-6194-4ae9-a7d1-1865a5be9583" containerName="ovn-config" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.454230 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9318e8c4-6194-4ae9-a7d1-1865a5be9583" containerName="ovn-config" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.454524 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9318e8c4-6194-4ae9-a7d1-1865a5be9583" containerName="ovn-config" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.455162 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.458508 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.470580 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qtj9x-config-fwvq8"] Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.568772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run-ovn\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.568832 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-log-ovn\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.569058 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.569125 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-additional-scripts\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.569147 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4s9r\" (UniqueName: \"kubernetes.io/projected/f841508d-0c8c-4856-a153-bc1bae588fcc-kube-api-access-s4s9r\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.569167 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-scripts\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.569588 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.670174 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-additional-scripts\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.670479 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4s9r\" (UniqueName: \"kubernetes.io/projected/f841508d-0c8c-4856-a153-bc1bae588fcc-kube-api-access-s4s9r\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.670594 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-scripts\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.671303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-additional-scripts\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.672606 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-scripts\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.673051 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run-ovn\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.673212 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-log-ovn\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.673314 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.673371 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run-ovn\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.673440 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-log-ovn\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.673561 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.696390 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4s9r\" (UniqueName: \"kubernetes.io/projected/f841508d-0c8c-4856-a153-bc1bae588fcc-kube-api-access-s4s9r\") pod \"ovn-controller-qtj9x-config-fwvq8\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:18 crc kubenswrapper[4810]: I1008 06:48:18.778395 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:19 crc kubenswrapper[4810]: I1008 06:48:19.153220 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:48:19 crc kubenswrapper[4810]: I1008 06:48:19.407073 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qtj9x-config-fwvq8"] Oct 08 06:48:19 crc kubenswrapper[4810]: I1008 06:48:19.861015 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x-config-fwvq8" event={"ID":"f841508d-0c8c-4856-a153-bc1bae588fcc","Type":"ContainerStarted","Data":"80569e9d82763f7df4441c225a43c04f1089ccfd570f2a0b75c3f7ec147cdb62"} Oct 08 06:48:19 crc kubenswrapper[4810]: I1008 06:48:19.861919 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x-config-fwvq8" event={"ID":"f841508d-0c8c-4856-a153-bc1bae588fcc","Type":"ContainerStarted","Data":"6fdbb2ec0bf5937c074a705b0b4afd9ea0bfe45e2baafc0af7b6c5b66f9216ae"} Oct 08 06:48:19 crc kubenswrapper[4810]: I1008 06:48:19.872501 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"f46dd8e3225e4f46200299b022317c9b2569c98b96714c24a69495ae35120418"} Oct 08 06:48:19 crc kubenswrapper[4810]: I1008 06:48:19.872568 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"e5382afb7813f143406f3f3d996682bc6f5be13125adec6966b7d947409db0b5"} Oct 08 06:48:19 crc kubenswrapper[4810]: I1008 06:48:19.872582 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"d8f77941e09530d2318f1aa89644df82f0915aa3cfe954e93ebcfc8a0be63897"} Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.084213 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9318e8c4-6194-4ae9-a7d1-1865a5be9583" path="/var/lib/kubelet/pods/9318e8c4-6194-4ae9-a7d1-1865a5be9583/volumes" Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.881400 4810 generic.go:334] "Generic (PLEG): container finished" podID="f841508d-0c8c-4856-a153-bc1bae588fcc" containerID="80569e9d82763f7df4441c225a43c04f1089ccfd570f2a0b75c3f7ec147cdb62" exitCode=0 Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.881591 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x-config-fwvq8" event={"ID":"f841508d-0c8c-4856-a153-bc1bae588fcc","Type":"ContainerDied","Data":"80569e9d82763f7df4441c225a43c04f1089ccfd570f2a0b75c3f7ec147cdb62"} Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.888116 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"758c08848dcf56803d3340df9307815df8fed52c268f9d576bd87dce38f15071"} Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.888151 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"b0eb359ab07ae0f5e3fe4ee340568b60987dd5232e6f155a7a05a079ef8c5e07"} Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.888163 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"345bb23b7033db196c378bf859a6c3b70c55475aa015ae3af5ecd4503e2dcf38"} Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.888173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerStarted","Data":"600f7e0d7b7425e861bb2335da4e0b3bf0e9e3610e33eb0340d0e575eba425da"} Oct 08 06:48:20 crc kubenswrapper[4810]: I1008 06:48:20.965285 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.103666542 podStartE2EDuration="26.965259612s" podCreationTimestamp="2025-10-08 06:47:54 +0000 UTC" firstStartedPulling="2025-10-08 06:48:12.112150856 +0000 UTC m=+994.746590596" lastFinishedPulling="2025-10-08 06:48:18.973743926 +0000 UTC m=+1001.608183666" observedRunningTime="2025-10-08 06:48:20.962806324 +0000 UTC m=+1003.597246074" watchObservedRunningTime="2025-10-08 06:48:20.965259612 +0000 UTC m=+1003.599699352" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.221412 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-564965cbfc-6q8bc"] Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.224641 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.226703 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.261033 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-564965cbfc-6q8bc"] Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.328057 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-config\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.328130 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-svc\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.328239 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nc6k\" (UniqueName: \"kubernetes.io/projected/78b447d3-3cd5-440a-8879-02c91e23e043-kube-api-access-8nc6k\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.328657 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-nb\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.328721 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-swift-storage-0\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.328838 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-sb\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.430848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-nb\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.430909 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-swift-storage-0\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.431019 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-sb\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.432303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-swift-storage-0\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.432368 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-config\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.432429 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-svc\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.432508 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-sb\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.433034 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-nb\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.433220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-svc\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.433343 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-config\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.433384 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nc6k\" (UniqueName: \"kubernetes.io/projected/78b447d3-3cd5-440a-8879-02c91e23e043-kube-api-access-8nc6k\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.457111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nc6k\" (UniqueName: \"kubernetes.io/projected/78b447d3-3cd5-440a-8879-02c91e23e043-kube-api-access-8nc6k\") pod \"dnsmasq-dns-564965cbfc-6q8bc\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:21 crc kubenswrapper[4810]: I1008 06:48:21.545549 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:27 crc kubenswrapper[4810]: I1008 06:48:27.987892 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x-config-fwvq8" event={"ID":"f841508d-0c8c-4856-a153-bc1bae588fcc","Type":"ContainerDied","Data":"6fdbb2ec0bf5937c074a705b0b4afd9ea0bfe45e2baafc0af7b6c5b66f9216ae"} Oct 08 06:48:27 crc kubenswrapper[4810]: I1008 06:48:27.988746 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fdbb2ec0bf5937c074a705b0b4afd9ea0bfe45e2baafc0af7b6c5b66f9216ae" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.100908 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.172912 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-scripts\") pod \"f841508d-0c8c-4856-a153-bc1bae588fcc\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.172987 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-log-ovn\") pod \"f841508d-0c8c-4856-a153-bc1bae588fcc\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.173042 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4s9r\" (UniqueName: \"kubernetes.io/projected/f841508d-0c8c-4856-a153-bc1bae588fcc-kube-api-access-s4s9r\") pod \"f841508d-0c8c-4856-a153-bc1bae588fcc\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.173176 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-additional-scripts\") pod \"f841508d-0c8c-4856-a153-bc1bae588fcc\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.173203 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run\") pod \"f841508d-0c8c-4856-a153-bc1bae588fcc\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.173289 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run-ovn\") pod \"f841508d-0c8c-4856-a153-bc1bae588fcc\" (UID: \"f841508d-0c8c-4856-a153-bc1bae588fcc\") " Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.173454 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run" (OuterVolumeSpecName: "var-run") pod "f841508d-0c8c-4856-a153-bc1bae588fcc" (UID: "f841508d-0c8c-4856-a153-bc1bae588fcc"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.173538 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f841508d-0c8c-4856-a153-bc1bae588fcc" (UID: "f841508d-0c8c-4856-a153-bc1bae588fcc"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.173643 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f841508d-0c8c-4856-a153-bc1bae588fcc" (UID: "f841508d-0c8c-4856-a153-bc1bae588fcc"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.174049 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f841508d-0c8c-4856-a153-bc1bae588fcc" (UID: "f841508d-0c8c-4856-a153-bc1bae588fcc"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.174424 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.174455 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.174466 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f841508d-0c8c-4856-a153-bc1bae588fcc-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.174874 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-scripts" (OuterVolumeSpecName: "scripts") pod "f841508d-0c8c-4856-a153-bc1bae588fcc" (UID: "f841508d-0c8c-4856-a153-bc1bae588fcc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.177699 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f841508d-0c8c-4856-a153-bc1bae588fcc-kube-api-access-s4s9r" (OuterVolumeSpecName: "kube-api-access-s4s9r") pod "f841508d-0c8c-4856-a153-bc1bae588fcc" (UID: "f841508d-0c8c-4856-a153-bc1bae588fcc"). InnerVolumeSpecName "kube-api-access-s4s9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.276515 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.276593 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4s9r\" (UniqueName: \"kubernetes.io/projected/f841508d-0c8c-4856-a153-bc1bae588fcc-kube-api-access-s4s9r\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.276606 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f841508d-0c8c-4856-a153-bc1bae588fcc-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.450830 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-564965cbfc-6q8bc"] Oct 08 06:48:28 crc kubenswrapper[4810]: W1008 06:48:28.465425 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78b447d3_3cd5_440a_8879_02c91e23e043.slice/crio-9c13e6604196f164be6eb62fd044882b7ac444369f7f4a64bb37fceb06ca79cb WatchSource:0}: Error finding container 9c13e6604196f164be6eb62fd044882b7ac444369f7f4a64bb37fceb06ca79cb: Status 404 returned error can't find the container with id 9c13e6604196f164be6eb62fd044882b7ac444369f7f4a64bb37fceb06ca79cb Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.567071 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.978467 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-h68hp"] Oct 08 06:48:28 crc kubenswrapper[4810]: E1008 06:48:28.979060 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f841508d-0c8c-4856-a153-bc1bae588fcc" containerName="ovn-config" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.979090 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f841508d-0c8c-4856-a153-bc1bae588fcc" containerName="ovn-config" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.979326 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f841508d-0c8c-4856-a153-bc1bae588fcc" containerName="ovn-config" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.980207 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h68hp" Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.988557 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-h68hp"] Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.998706 4810 generic.go:334] "Generic (PLEG): container finished" podID="78b447d3-3cd5-440a-8879-02c91e23e043" containerID="6cd4b9c52d92efd669812c350f816e6f8d083bdac9339adf49e3105f6eb37d82" exitCode=0 Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.998797 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" event={"ID":"78b447d3-3cd5-440a-8879-02c91e23e043","Type":"ContainerDied","Data":"6cd4b9c52d92efd669812c350f816e6f8d083bdac9339adf49e3105f6eb37d82"} Oct 08 06:48:28 crc kubenswrapper[4810]: I1008 06:48:28.998836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" event={"ID":"78b447d3-3cd5-440a-8879-02c91e23e043","Type":"ContainerStarted","Data":"9c13e6604196f164be6eb62fd044882b7ac444369f7f4a64bb37fceb06ca79cb"} Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.001331 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x-config-fwvq8" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.001413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lcpxn" event={"ID":"c40f9ad6-fed6-4759-8aa3-134912fd46bc","Type":"ContainerStarted","Data":"cb62f083c1a6339ca4b73ea3bc19a767cc24f149de31aa454da4c2003b472a7c"} Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.080665 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-lcpxn" podStartSLOduration=3.782460131 podStartE2EDuration="17.080645099s" podCreationTimestamp="2025-10-08 06:48:12 +0000 UTC" firstStartedPulling="2025-10-08 06:48:14.674766147 +0000 UTC m=+997.309205887" lastFinishedPulling="2025-10-08 06:48:27.972951105 +0000 UTC m=+1010.607390855" observedRunningTime="2025-10-08 06:48:29.074641133 +0000 UTC m=+1011.709080873" watchObservedRunningTime="2025-10-08 06:48:29.080645099 +0000 UTC m=+1011.715084839" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.094438 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shc7d\" (UniqueName: \"kubernetes.io/projected/a29b98cd-a057-457e-91f5-aa6e7a05b840-kube-api-access-shc7d\") pod \"cinder-db-create-h68hp\" (UID: \"a29b98cd-a057-457e-91f5-aa6e7a05b840\") " pod="openstack/cinder-db-create-h68hp" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.109615 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-g4kvt"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.111085 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g4kvt" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.122173 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-g4kvt"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.206558 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqvc4\" (UniqueName: \"kubernetes.io/projected/bc50ffeb-5f89-4a44-a95f-2321d36f6fb7-kube-api-access-mqvc4\") pod \"barbican-db-create-g4kvt\" (UID: \"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7\") " pod="openstack/barbican-db-create-g4kvt" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.206726 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shc7d\" (UniqueName: \"kubernetes.io/projected/a29b98cd-a057-457e-91f5-aa6e7a05b840-kube-api-access-shc7d\") pod \"cinder-db-create-h68hp\" (UID: \"a29b98cd-a057-457e-91f5-aa6e7a05b840\") " pod="openstack/cinder-db-create-h68hp" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.222215 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-qtj9x-config-fwvq8"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.226652 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-qtj9x-config-fwvq8"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.228318 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shc7d\" (UniqueName: \"kubernetes.io/projected/a29b98cd-a057-457e-91f5-aa6e7a05b840-kube-api-access-shc7d\") pod \"cinder-db-create-h68hp\" (UID: \"a29b98cd-a057-457e-91f5-aa6e7a05b840\") " pod="openstack/cinder-db-create-h68hp" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.263990 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-5qq97"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.265683 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.268712 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.269016 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.269171 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.269325 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tb5ft" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.281274 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-5qq97"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.297905 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h68hp" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.315152 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqvc4\" (UniqueName: \"kubernetes.io/projected/bc50ffeb-5f89-4a44-a95f-2321d36f6fb7-kube-api-access-mqvc4\") pod \"barbican-db-create-g4kvt\" (UID: \"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7\") " pod="openstack/barbican-db-create-g4kvt" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.342365 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqvc4\" (UniqueName: \"kubernetes.io/projected/bc50ffeb-5f89-4a44-a95f-2321d36f6fb7-kube-api-access-mqvc4\") pod \"barbican-db-create-g4kvt\" (UID: \"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7\") " pod="openstack/barbican-db-create-g4kvt" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.370400 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-4r7z7"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.371409 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4r7z7" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.421879 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-combined-ca-bundle\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.422140 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8lgz\" (UniqueName: \"kubernetes.io/projected/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-kube-api-access-j8lgz\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.422246 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-config-data\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.424126 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4r7z7"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.495205 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g4kvt" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.523719 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-combined-ca-bundle\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.523902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8lgz\" (UniqueName: \"kubernetes.io/projected/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-kube-api-access-j8lgz\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.524030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw4j7\" (UniqueName: \"kubernetes.io/projected/d6755119-3241-4b88-b417-a13caf382092-kube-api-access-qw4j7\") pod \"neutron-db-create-4r7z7\" (UID: \"d6755119-3241-4b88-b417-a13caf382092\") " pod="openstack/neutron-db-create-4r7z7" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.524137 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-config-data\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.530426 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-combined-ca-bundle\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.531931 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-config-data\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.546829 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8lgz\" (UniqueName: \"kubernetes.io/projected/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-kube-api-access-j8lgz\") pod \"keystone-db-sync-5qq97\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.588218 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.625462 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw4j7\" (UniqueName: \"kubernetes.io/projected/d6755119-3241-4b88-b417-a13caf382092-kube-api-access-qw4j7\") pod \"neutron-db-create-4r7z7\" (UID: \"d6755119-3241-4b88-b417-a13caf382092\") " pod="openstack/neutron-db-create-4r7z7" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.647816 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw4j7\" (UniqueName: \"kubernetes.io/projected/d6755119-3241-4b88-b417-a13caf382092-kube-api-access-qw4j7\") pod \"neutron-db-create-4r7z7\" (UID: \"d6755119-3241-4b88-b417-a13caf382092\") " pod="openstack/neutron-db-create-4r7z7" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.691643 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4r7z7" Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.847084 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-h68hp"] Oct 08 06:48:29 crc kubenswrapper[4810]: I1008 06:48:29.943372 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-5qq97"] Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.017618 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" event={"ID":"78b447d3-3cd5-440a-8879-02c91e23e043","Type":"ContainerStarted","Data":"b7c833b41867e5b68178094e9963af557ce11cc0418b0f74d99fcdb749a941cc"} Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.018861 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.020636 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-h68hp" event={"ID":"a29b98cd-a057-457e-91f5-aa6e7a05b840","Type":"ContainerStarted","Data":"7de2ef987c8118efde64f5b974a65a102e4b6390380d11b871bd6043895cdc1f"} Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.023537 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5qq97" event={"ID":"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477","Type":"ContainerStarted","Data":"f42c370ca6e7e73a5c23a69d22269f7ae72f982b80ae07f432852e652724b716"} Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.026537 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-g4kvt"] Oct 08 06:48:30 crc kubenswrapper[4810]: W1008 06:48:30.034527 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc50ffeb_5f89_4a44_a95f_2321d36f6fb7.slice/crio-679fce53af2545b9c2e401ed94b453bad5922a77f1af3d2981a3b0aa65b0f0c4 WatchSource:0}: Error finding container 679fce53af2545b9c2e401ed94b453bad5922a77f1af3d2981a3b0aa65b0f0c4: Status 404 returned error can't find the container with id 679fce53af2545b9c2e401ed94b453bad5922a77f1af3d2981a3b0aa65b0f0c4 Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.049399 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" podStartSLOduration=9.049377408 podStartE2EDuration="9.049377408s" podCreationTimestamp="2025-10-08 06:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:30.038772673 +0000 UTC m=+1012.673212423" watchObservedRunningTime="2025-10-08 06:48:30.049377408 +0000 UTC m=+1012.683817148" Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.094165 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f841508d-0c8c-4856-a153-bc1bae588fcc" path="/var/lib/kubelet/pods/f841508d-0c8c-4856-a153-bc1bae588fcc/volumes" Oct 08 06:48:30 crc kubenswrapper[4810]: I1008 06:48:30.241950 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4r7z7"] Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.047212 4810 generic.go:334] "Generic (PLEG): container finished" podID="d6755119-3241-4b88-b417-a13caf382092" containerID="f60d12361f4468b41bc7cb57d821f3097194d727ab08709eb1949abd417c69a2" exitCode=0 Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.047639 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4r7z7" event={"ID":"d6755119-3241-4b88-b417-a13caf382092","Type":"ContainerDied","Data":"f60d12361f4468b41bc7cb57d821f3097194d727ab08709eb1949abd417c69a2"} Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.047671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4r7z7" event={"ID":"d6755119-3241-4b88-b417-a13caf382092","Type":"ContainerStarted","Data":"4fa4f1b2abc70bac4ee805c42c267caa660e274010f441eeb31d0110cdc46c3f"} Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.049721 4810 generic.go:334] "Generic (PLEG): container finished" podID="a29b98cd-a057-457e-91f5-aa6e7a05b840" containerID="eacb2363992ea55cecd62e4498198b8367c189671343114cf21204d0ef93a535" exitCode=0 Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.049779 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-h68hp" event={"ID":"a29b98cd-a057-457e-91f5-aa6e7a05b840","Type":"ContainerDied","Data":"eacb2363992ea55cecd62e4498198b8367c189671343114cf21204d0ef93a535"} Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.051292 4810 generic.go:334] "Generic (PLEG): container finished" podID="bc50ffeb-5f89-4a44-a95f-2321d36f6fb7" containerID="e5867c8c5052a231dc51cd45cde9fddb5279855338c65d3dc46514e20fd68fcf" exitCode=0 Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.052079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g4kvt" event={"ID":"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7","Type":"ContainerDied","Data":"e5867c8c5052a231dc51cd45cde9fddb5279855338c65d3dc46514e20fd68fcf"} Oct 08 06:48:31 crc kubenswrapper[4810]: I1008 06:48:31.052103 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g4kvt" event={"ID":"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7","Type":"ContainerStarted","Data":"679fce53af2545b9c2e401ed94b453bad5922a77f1af3d2981a3b0aa65b0f0c4"} Oct 08 06:48:36 crc kubenswrapper[4810]: I1008 06:48:36.548209 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:36 crc kubenswrapper[4810]: I1008 06:48:36.630357 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-xqcwk"] Oct 08 06:48:36 crc kubenswrapper[4810]: I1008 06:48:36.631236 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerName="dnsmasq-dns" containerID="cri-o://5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348" gracePeriod=10 Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.341443 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h68hp" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.444495 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g4kvt" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.475641 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4r7z7" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.490204 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.524554 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shc7d\" (UniqueName: \"kubernetes.io/projected/a29b98cd-a057-457e-91f5-aa6e7a05b840-kube-api-access-shc7d\") pod \"a29b98cd-a057-457e-91f5-aa6e7a05b840\" (UID: \"a29b98cd-a057-457e-91f5-aa6e7a05b840\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.532673 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29b98cd-a057-457e-91f5-aa6e7a05b840-kube-api-access-shc7d" (OuterVolumeSpecName: "kube-api-access-shc7d") pod "a29b98cd-a057-457e-91f5-aa6e7a05b840" (UID: "a29b98cd-a057-457e-91f5-aa6e7a05b840"). InnerVolumeSpecName "kube-api-access-shc7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626265 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-dns-svc\") pod \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626323 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8xvn\" (UniqueName: \"kubernetes.io/projected/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-kube-api-access-p8xvn\") pod \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626354 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-nb\") pod \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626393 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw4j7\" (UniqueName: \"kubernetes.io/projected/d6755119-3241-4b88-b417-a13caf382092-kube-api-access-qw4j7\") pod \"d6755119-3241-4b88-b417-a13caf382092\" (UID: \"d6755119-3241-4b88-b417-a13caf382092\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626449 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqvc4\" (UniqueName: \"kubernetes.io/projected/bc50ffeb-5f89-4a44-a95f-2321d36f6fb7-kube-api-access-mqvc4\") pod \"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7\" (UID: \"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626465 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-config\") pod \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626517 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-sb\") pod \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\" (UID: \"b3b1c5e5-07c2-4c7e-9a71-bd427886615f\") " Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.626849 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shc7d\" (UniqueName: \"kubernetes.io/projected/a29b98cd-a057-457e-91f5-aa6e7a05b840-kube-api-access-shc7d\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.630983 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc50ffeb-5f89-4a44-a95f-2321d36f6fb7-kube-api-access-mqvc4" (OuterVolumeSpecName: "kube-api-access-mqvc4") pod "bc50ffeb-5f89-4a44-a95f-2321d36f6fb7" (UID: "bc50ffeb-5f89-4a44-a95f-2321d36f6fb7"). InnerVolumeSpecName "kube-api-access-mqvc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.632411 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-kube-api-access-p8xvn" (OuterVolumeSpecName: "kube-api-access-p8xvn") pod "b3b1c5e5-07c2-4c7e-9a71-bd427886615f" (UID: "b3b1c5e5-07c2-4c7e-9a71-bd427886615f"). InnerVolumeSpecName "kube-api-access-p8xvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.635092 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6755119-3241-4b88-b417-a13caf382092-kube-api-access-qw4j7" (OuterVolumeSpecName: "kube-api-access-qw4j7") pod "d6755119-3241-4b88-b417-a13caf382092" (UID: "d6755119-3241-4b88-b417-a13caf382092"). InnerVolumeSpecName "kube-api-access-qw4j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.671558 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3b1c5e5-07c2-4c7e-9a71-bd427886615f" (UID: "b3b1c5e5-07c2-4c7e-9a71-bd427886615f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.685258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3b1c5e5-07c2-4c7e-9a71-bd427886615f" (UID: "b3b1c5e5-07c2-4c7e-9a71-bd427886615f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.691578 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-config" (OuterVolumeSpecName: "config") pod "b3b1c5e5-07c2-4c7e-9a71-bd427886615f" (UID: "b3b1c5e5-07c2-4c7e-9a71-bd427886615f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.697894 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3b1c5e5-07c2-4c7e-9a71-bd427886615f" (UID: "b3b1c5e5-07c2-4c7e-9a71-bd427886615f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.728867 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.728910 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8xvn\" (UniqueName: \"kubernetes.io/projected/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-kube-api-access-p8xvn\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.728923 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.728933 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw4j7\" (UniqueName: \"kubernetes.io/projected/d6755119-3241-4b88-b417-a13caf382092-kube-api-access-qw4j7\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.728943 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqvc4\" (UniqueName: \"kubernetes.io/projected/bc50ffeb-5f89-4a44-a95f-2321d36f6fb7-kube-api-access-mqvc4\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.728952 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:39 crc kubenswrapper[4810]: I1008 06:48:39.728975 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3b1c5e5-07c2-4c7e-9a71-bd427886615f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.152765 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g4kvt" event={"ID":"bc50ffeb-5f89-4a44-a95f-2321d36f6fb7","Type":"ContainerDied","Data":"679fce53af2545b9c2e401ed94b453bad5922a77f1af3d2981a3b0aa65b0f0c4"} Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.152805 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="679fce53af2545b9c2e401ed94b453bad5922a77f1af3d2981a3b0aa65b0f0c4" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.152871 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g4kvt" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.159608 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5qq97" event={"ID":"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477","Type":"ContainerStarted","Data":"28edf5eb737efa28788525bfeb7f9c64a807fad3bc80060ae62756f6fad80ae2"} Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.162299 4810 generic.go:334] "Generic (PLEG): container finished" podID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerID="5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348" exitCode=0 Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.162339 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" event={"ID":"b3b1c5e5-07c2-4c7e-9a71-bd427886615f","Type":"ContainerDied","Data":"5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348"} Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.162356 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" event={"ID":"b3b1c5e5-07c2-4c7e-9a71-bd427886615f","Type":"ContainerDied","Data":"0ecd67f7f73d631488f5153082c753ebb5d4e10ce4146f0818533dce00cd5bb7"} Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.162372 4810 scope.go:117] "RemoveContainer" containerID="5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.162464 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.173081 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-h68hp" event={"ID":"a29b98cd-a057-457e-91f5-aa6e7a05b840","Type":"ContainerDied","Data":"7de2ef987c8118efde64f5b974a65a102e4b6390380d11b871bd6043895cdc1f"} Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.173133 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7de2ef987c8118efde64f5b974a65a102e4b6390380d11b871bd6043895cdc1f" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.173260 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-h68hp" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.178107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4r7z7" event={"ID":"d6755119-3241-4b88-b417-a13caf382092","Type":"ContainerDied","Data":"4fa4f1b2abc70bac4ee805c42c267caa660e274010f441eeb31d0110cdc46c3f"} Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.178227 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fa4f1b2abc70bac4ee805c42c267caa660e274010f441eeb31d0110cdc46c3f" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.178188 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4r7z7" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.190911 4810 scope.go:117] "RemoveContainer" containerID="7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.194181 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-xqcwk"] Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.208107 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-xqcwk"] Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.208894 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-5qq97" podStartSLOduration=1.953616016 podStartE2EDuration="11.208875409s" podCreationTimestamp="2025-10-08 06:48:29 +0000 UTC" firstStartedPulling="2025-10-08 06:48:29.97233695 +0000 UTC m=+1012.606776680" lastFinishedPulling="2025-10-08 06:48:39.227596313 +0000 UTC m=+1021.862036073" observedRunningTime="2025-10-08 06:48:40.206549205 +0000 UTC m=+1022.840988965" watchObservedRunningTime="2025-10-08 06:48:40.208875409 +0000 UTC m=+1022.843315149" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.223331 4810 scope.go:117] "RemoveContainer" containerID="5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348" Oct 08 06:48:40 crc kubenswrapper[4810]: E1008 06:48:40.223877 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348\": container with ID starting with 5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348 not found: ID does not exist" containerID="5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.223921 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348"} err="failed to get container status \"5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348\": rpc error: code = NotFound desc = could not find container \"5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348\": container with ID starting with 5701b485f88aadec071b2e5c81a433de801a399c0a4fa2185c034e9fa1fba348 not found: ID does not exist" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.223948 4810 scope.go:117] "RemoveContainer" containerID="7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32" Oct 08 06:48:40 crc kubenswrapper[4810]: E1008 06:48:40.224223 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32\": container with ID starting with 7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32 not found: ID does not exist" containerID="7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32" Oct 08 06:48:40 crc kubenswrapper[4810]: I1008 06:48:40.224249 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32"} err="failed to get container status \"7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32\": rpc error: code = NotFound desc = could not find container \"7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32\": container with ID starting with 7610ed58e613f2d49d9850aaf7aa6f25f52dbb6c049db6fb203f76c3ba3d2b32 not found: ID does not exist" Oct 08 06:48:41 crc kubenswrapper[4810]: I1008 06:48:41.190816 4810 generic.go:334] "Generic (PLEG): container finished" podID="c40f9ad6-fed6-4759-8aa3-134912fd46bc" containerID="cb62f083c1a6339ca4b73ea3bc19a767cc24f149de31aa454da4c2003b472a7c" exitCode=0 Oct 08 06:48:41 crc kubenswrapper[4810]: I1008 06:48:41.190892 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lcpxn" event={"ID":"c40f9ad6-fed6-4759-8aa3-134912fd46bc","Type":"ContainerDied","Data":"cb62f083c1a6339ca4b73ea3bc19a767cc24f149de31aa454da4c2003b472a7c"} Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.094227 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" path="/var/lib/kubelet/pods/b3b1c5e5-07c2-4c7e-9a71-bd427886615f/volumes" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.205607 4810 generic.go:334] "Generic (PLEG): container finished" podID="5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" containerID="28edf5eb737efa28788525bfeb7f9c64a807fad3bc80060ae62756f6fad80ae2" exitCode=0 Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.205692 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5qq97" event={"ID":"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477","Type":"ContainerDied","Data":"28edf5eb737efa28788525bfeb7f9c64a807fad3bc80060ae62756f6fad80ae2"} Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.651097 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.778258 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-config-data\") pod \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.778482 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-db-sync-config-data\") pod \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.778538 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-combined-ca-bundle\") pod \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.778579 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9zwd\" (UniqueName: \"kubernetes.io/projected/c40f9ad6-fed6-4759-8aa3-134912fd46bc-kube-api-access-n9zwd\") pod \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\" (UID: \"c40f9ad6-fed6-4759-8aa3-134912fd46bc\") " Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.786005 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c40f9ad6-fed6-4759-8aa3-134912fd46bc" (UID: "c40f9ad6-fed6-4759-8aa3-134912fd46bc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.786061 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c40f9ad6-fed6-4759-8aa3-134912fd46bc-kube-api-access-n9zwd" (OuterVolumeSpecName: "kube-api-access-n9zwd") pod "c40f9ad6-fed6-4759-8aa3-134912fd46bc" (UID: "c40f9ad6-fed6-4759-8aa3-134912fd46bc"). InnerVolumeSpecName "kube-api-access-n9zwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.810581 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c40f9ad6-fed6-4759-8aa3-134912fd46bc" (UID: "c40f9ad6-fed6-4759-8aa3-134912fd46bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.829219 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-config-data" (OuterVolumeSpecName: "config-data") pod "c40f9ad6-fed6-4759-8aa3-134912fd46bc" (UID: "c40f9ad6-fed6-4759-8aa3-134912fd46bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.880321 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9zwd\" (UniqueName: \"kubernetes.io/projected/c40f9ad6-fed6-4759-8aa3-134912fd46bc-kube-api-access-n9zwd\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.880364 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.880374 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:42 crc kubenswrapper[4810]: I1008 06:48:42.880382 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c40f9ad6-fed6-4759-8aa3-134912fd46bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.233502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lcpxn" event={"ID":"c40f9ad6-fed6-4759-8aa3-134912fd46bc","Type":"ContainerDied","Data":"ebc0e69e8e304e5e88de2824441e3f694967b5bc722a024b0dde1d9ec6097c13"} Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.233555 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebc0e69e8e304e5e88de2824441e3f694967b5bc722a024b0dde1d9ec6097c13" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.233521 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lcpxn" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.629739 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-795846498c-pp6hm"] Oct 08 06:48:43 crc kubenswrapper[4810]: E1008 06:48:43.631041 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerName="init" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631065 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerName="init" Oct 08 06:48:43 crc kubenswrapper[4810]: E1008 06:48:43.631105 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40f9ad6-fed6-4759-8aa3-134912fd46bc" containerName="glance-db-sync" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631114 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40f9ad6-fed6-4759-8aa3-134912fd46bc" containerName="glance-db-sync" Oct 08 06:48:43 crc kubenswrapper[4810]: E1008 06:48:43.631138 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc50ffeb-5f89-4a44-a95f-2321d36f6fb7" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631146 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc50ffeb-5f89-4a44-a95f-2321d36f6fb7" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: E1008 06:48:43.631173 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29b98cd-a057-457e-91f5-aa6e7a05b840" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631182 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29b98cd-a057-457e-91f5-aa6e7a05b840" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: E1008 06:48:43.631222 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerName="dnsmasq-dns" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631230 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerName="dnsmasq-dns" Oct 08 06:48:43 crc kubenswrapper[4810]: E1008 06:48:43.631245 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6755119-3241-4b88-b417-a13caf382092" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631252 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6755119-3241-4b88-b417-a13caf382092" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631669 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a29b98cd-a057-457e-91f5-aa6e7a05b840" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631705 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40f9ad6-fed6-4759-8aa3-134912fd46bc" containerName="glance-db-sync" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631733 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerName="dnsmasq-dns" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631748 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6755119-3241-4b88-b417-a13caf382092" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.631774 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc50ffeb-5f89-4a44-a95f-2321d36f6fb7" containerName="mariadb-database-create" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.634082 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.665990 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.701624 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795846498c-pp6hm"] Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702053 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-config-data\") pod \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702102 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8lgz\" (UniqueName: \"kubernetes.io/projected/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-kube-api-access-j8lgz\") pod \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702199 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-combined-ca-bundle\") pod \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\" (UID: \"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477\") " Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702355 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-sb\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702423 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-config\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702473 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-swift-storage-0\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702521 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-svc\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702556 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-nb\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.702654 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwzxc\" (UniqueName: \"kubernetes.io/projected/f866fb32-3995-422b-a267-e0cc005d9c12-kube-api-access-fwzxc\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.731257 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-kube-api-access-j8lgz" (OuterVolumeSpecName: "kube-api-access-j8lgz") pod "5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" (UID: "5afd09d8-444a-4cc9-a25c-fe3a0b0d8477"). InnerVolumeSpecName "kube-api-access-j8lgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.744329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" (UID: "5afd09d8-444a-4cc9-a25c-fe3a0b0d8477"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.784613 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-config-data" (OuterVolumeSpecName: "config-data") pod "5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" (UID: "5afd09d8-444a-4cc9-a25c-fe3a0b0d8477"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.808812 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-sb\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.808880 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-config\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.808919 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-swift-storage-0\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.808992 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-svc\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.809023 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-nb\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.809077 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwzxc\" (UniqueName: \"kubernetes.io/projected/f866fb32-3995-422b-a267-e0cc005d9c12-kube-api-access-fwzxc\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.809146 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.809159 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8lgz\" (UniqueName: \"kubernetes.io/projected/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-kube-api-access-j8lgz\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.809171 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.811148 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-svc\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.811176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-nb\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.811140 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-swift-storage-0\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.811792 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-config\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.811838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-sb\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:43 crc kubenswrapper[4810]: I1008 06:48:43.831944 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwzxc\" (UniqueName: \"kubernetes.io/projected/f866fb32-3995-422b-a267-e0cc005d9c12-kube-api-access-fwzxc\") pod \"dnsmasq-dns-795846498c-pp6hm\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.001590 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.246433 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-5qq97" event={"ID":"5afd09d8-444a-4cc9-a25c-fe3a0b0d8477","Type":"ContainerDied","Data":"f42c370ca6e7e73a5c23a69d22269f7ae72f982b80ae07f432852e652724b716"} Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.246812 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f42c370ca6e7e73a5c23a69d22269f7ae72f982b80ae07f432852e652724b716" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.246475 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-5qq97" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.381063 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795846498c-pp6hm"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.414211 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2x6lv"] Oct 08 06:48:44 crc kubenswrapper[4810]: E1008 06:48:44.414649 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" containerName="keystone-db-sync" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.414671 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" containerName="keystone-db-sync" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.414893 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" containerName="keystone-db-sync" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.415546 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.418598 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.418753 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.418788 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.418881 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tb5ft" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.430518 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b4bfdd7f7-9m56r"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.432209 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7b587f8db7-xqcwk" podUID="b3b1c5e5-07c2-4c7e-9a71-bd427886615f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: i/o timeout" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.432625 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.476450 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2x6lv"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.484974 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b4bfdd7f7-9m56r"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.521819 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-fernet-keys\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.521870 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpg5b\" (UniqueName: \"kubernetes.io/projected/0650b229-1dc2-454b-b2b5-62adc442b564-kube-api-access-cpg5b\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.521902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-config-data\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.521990 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-combined-ca-bundle\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.522012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-credential-keys\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.522031 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-scripts\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: W1008 06:48:44.554444 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf866fb32_3995_422b_a267_e0cc005d9c12.slice/crio-8f023d7ba39594a3f04d18e9eb23dee8c24c749920b09c1301d09ab133ad5342 WatchSource:0}: Error finding container 8f023d7ba39594a3f04d18e9eb23dee8c24c749920b09c1301d09ab133ad5342: Status 404 returned error can't find the container with id 8f023d7ba39594a3f04d18e9eb23dee8c24c749920b09c1301d09ab133ad5342 Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.559324 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795846498c-pp6hm"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.623845 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltbp5\" (UniqueName: \"kubernetes.io/projected/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-kube-api-access-ltbp5\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-nb\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624304 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-fernet-keys\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpg5b\" (UniqueName: \"kubernetes.io/projected/0650b229-1dc2-454b-b2b5-62adc442b564-kube-api-access-cpg5b\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624358 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-config-data\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624385 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-config\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624444 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-sb\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624467 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-combined-ca-bundle\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624486 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-swift-storage-0\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624509 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-credential-keys\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-scripts\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.624554 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-svc\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.632645 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-credential-keys\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.633408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-fernet-keys\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.637379 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-config-data\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.640144 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-combined-ca-bundle\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.654187 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-scripts\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.673068 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpg5b\" (UniqueName: \"kubernetes.io/projected/0650b229-1dc2-454b-b2b5-62adc442b564-kube-api-access-cpg5b\") pod \"keystone-bootstrap-2x6lv\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.726369 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-svc\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.726629 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltbp5\" (UniqueName: \"kubernetes.io/projected/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-kube-api-access-ltbp5\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.726698 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-nb\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.726794 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-config\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.726935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-sb\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.727074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-swift-storage-0\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.727987 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-swift-storage-0\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.728672 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-svc\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.729369 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-nb\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.729668 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-config\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.730867 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-sb\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.731893 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b4bfdd7f7-9m56r"] Oct 08 06:48:44 crc kubenswrapper[4810]: E1008 06:48:44.737938 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-ltbp5], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" podUID="6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.761864 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltbp5\" (UniqueName: \"kubernetes.io/projected/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-kube-api-access-ltbp5\") pod \"dnsmasq-dns-6b4bfdd7f7-9m56r\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.782032 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dc68bd5-mpslw"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.784007 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.787514 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.787869 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.803144 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.811002 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.811261 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.820019 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dc68bd5-mpslw"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.836292 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.892288 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-h7z72"] Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.893458 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.907737 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qccns" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.914041 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.914276 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936062 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc5xm\" (UniqueName: \"kubernetes.io/projected/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-kube-api-access-nc5xm\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936157 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-swift-storage-0\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936242 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-config-data\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-nb\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936332 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-log-httpd\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-svc\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936470 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-sb\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936507 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-config\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936555 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-scripts\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936615 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-run-httpd\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936665 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gbl6\" (UniqueName: \"kubernetes.io/projected/9d5fc753-8316-4e77-8e4a-c0ab288b6589-kube-api-access-2gbl6\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936710 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.936767 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:44 crc kubenswrapper[4810]: I1008 06:48:44.945470 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-h7z72"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.038605 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-scripts\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.038858 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnkvs\" (UniqueName: \"kubernetes.io/projected/22742881-9836-4dbb-bd8c-deba4194d7c3-kube-api-access-qnkvs\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.038877 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-config-data\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.038903 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-sb\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.038920 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-config\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.038940 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-scripts\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.038981 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-run-httpd\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039000 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gbl6\" (UniqueName: \"kubernetes.io/projected/9d5fc753-8316-4e77-8e4a-c0ab288b6589-kube-api-access-2gbl6\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039019 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039043 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc5xm\" (UniqueName: \"kubernetes.io/projected/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-kube-api-access-nc5xm\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039090 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22742881-9836-4dbb-bd8c-deba4194d7c3-logs\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039107 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-combined-ca-bundle\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039129 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-swift-storage-0\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039160 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-config-data\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039184 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-nb\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039203 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-log-httpd\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.039223 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-svc\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.040405 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-sb\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.043647 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-run-httpd\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.044287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-log-httpd\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.050801 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-config-data\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.053216 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-config\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.054937 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-svc\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.056456 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-nb\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.057453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-swift-storage-0\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.062221 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.064642 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-scripts\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.080598 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.087630 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gbl6\" (UniqueName: \"kubernetes.io/projected/9d5fc753-8316-4e77-8e4a-c0ab288b6589-kube-api-access-2gbl6\") pod \"dnsmasq-dns-5dc68bd5-mpslw\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.088985 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc5xm\" (UniqueName: \"kubernetes.io/projected/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-kube-api-access-nc5xm\") pod \"ceilometer-0\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.121329 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.139076 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.140410 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-scripts\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.140526 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnkvs\" (UniqueName: \"kubernetes.io/projected/22742881-9836-4dbb-bd8c-deba4194d7c3-kube-api-access-qnkvs\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.140592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-config-data\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.140721 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22742881-9836-4dbb-bd8c-deba4194d7c3-logs\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.140787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-combined-ca-bundle\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.143797 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22742881-9836-4dbb-bd8c-deba4194d7c3-logs\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.151613 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-scripts\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.152314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-config-data\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.170316 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-combined-ca-bundle\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.182493 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnkvs\" (UniqueName: \"kubernetes.io/projected/22742881-9836-4dbb-bd8c-deba4194d7c3-kube-api-access-qnkvs\") pod \"placement-db-sync-h7z72\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.238839 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.287133 4810 generic.go:334] "Generic (PLEG): container finished" podID="f866fb32-3995-422b-a267-e0cc005d9c12" containerID="6faccb5e7c8065d98bfa6fafd38dd3b7eed39278107ded21bca68560f68b9cc3" exitCode=0 Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.288352 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.287388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795846498c-pp6hm" event={"ID":"f866fb32-3995-422b-a267-e0cc005d9c12","Type":"ContainerDied","Data":"6faccb5e7c8065d98bfa6fafd38dd3b7eed39278107ded21bca68560f68b9cc3"} Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.289397 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795846498c-pp6hm" event={"ID":"f866fb32-3995-422b-a267-e0cc005d9c12","Type":"ContainerStarted","Data":"8f023d7ba39594a3f04d18e9eb23dee8c24c749920b09c1301d09ab133ad5342"} Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.320481 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.443841 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2x6lv"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.449729 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-nb\") pod \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.449896 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-swift-storage-0\") pod \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.449934 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-config\") pod \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.450029 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-sb\") pod \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.450061 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-svc\") pod \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.450110 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltbp5\" (UniqueName: \"kubernetes.io/projected/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-kube-api-access-ltbp5\") pod \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\" (UID: \"6f48bbd0-a1c0-49ae-a133-d1c78b4869f9\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.451557 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-config" (OuterVolumeSpecName: "config") pod "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" (UID: "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.452050 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" (UID: "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.452340 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" (UID: "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.452371 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" (UID: "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.452545 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" (UID: "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.454747 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-kube-api-access-ltbp5" (OuterVolumeSpecName: "kube-api-access-ltbp5") pod "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" (UID: "6f48bbd0-a1c0-49ae-a133-d1c78b4869f9"). InnerVolumeSpecName "kube-api-access-ltbp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.553410 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.553441 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.553451 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.553459 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.553467 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltbp5\" (UniqueName: \"kubernetes.io/projected/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-kube-api-access-ltbp5\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.553476 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.602973 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.609396 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.615052 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.618265 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6dmpd" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.618400 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.618549 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.711592 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.727848 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.729604 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.732413 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.749381 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.758153 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-logs\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.758207 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.758256 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.758597 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6st2w\" (UniqueName: \"kubernetes.io/projected/6f7195c4-3f0e-412c-bf83-2862e38a22d5-kube-api-access-6st2w\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.758677 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-config-data\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.758748 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.758772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-scripts\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.795292 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.861672 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-logs\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.861740 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6st2w\" (UniqueName: \"kubernetes.io/projected/6f7195c4-3f0e-412c-bf83-2862e38a22d5-kube-api-access-6st2w\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.861860 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdpz2\" (UniqueName: \"kubernetes.io/projected/e413986a-213e-41e2-ab3e-2d2e8899110a-kube-api-access-pdpz2\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.861919 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-config-data\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862010 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862034 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862052 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862086 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-scripts\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862124 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-logs\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862167 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862249 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862304 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862336 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.862926 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.863225 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-logs\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.863863 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.865571 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dc68bd5-mpslw"] Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.875605 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-config-data\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.876107 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-scripts\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: W1008 06:48:45.876754 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d5fc753_8316_4e77_8e4a_c0ab288b6589.slice/crio-0f50f021b43579d636062d4ec7e244b3fcf631eb1d2bf79db6e4ec97b62f2eca WatchSource:0}: Error finding container 0f50f021b43579d636062d4ec7e244b3fcf631eb1d2bf79db6e4ec97b62f2eca: Status 404 returned error can't find the container with id 0f50f021b43579d636062d4ec7e244b3fcf631eb1d2bf79db6e4ec97b62f2eca Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.879399 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.885800 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6st2w\" (UniqueName: \"kubernetes.io/projected/6f7195c4-3f0e-412c-bf83-2862e38a22d5-kube-api-access-6st2w\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.897669 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.963726 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-svc\") pod \"f866fb32-3995-422b-a267-e0cc005d9c12\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964244 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-nb\") pod \"f866fb32-3995-422b-a267-e0cc005d9c12\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964317 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-config\") pod \"f866fb32-3995-422b-a267-e0cc005d9c12\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964350 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwzxc\" (UniqueName: \"kubernetes.io/projected/f866fb32-3995-422b-a267-e0cc005d9c12-kube-api-access-fwzxc\") pod \"f866fb32-3995-422b-a267-e0cc005d9c12\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964391 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-sb\") pod \"f866fb32-3995-422b-a267-e0cc005d9c12\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964418 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-swift-storage-0\") pod \"f866fb32-3995-422b-a267-e0cc005d9c12\" (UID: \"f866fb32-3995-422b-a267-e0cc005d9c12\") " Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964660 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964694 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964760 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964776 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964807 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-logs\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.964915 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdpz2\" (UniqueName: \"kubernetes.io/projected/e413986a-213e-41e2-ab3e-2d2e8899110a-kube-api-access-pdpz2\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.969476 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.970785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.971327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-logs\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.976022 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.979202 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.990645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f866fb32-3995-422b-a267-e0cc005d9c12-kube-api-access-fwzxc" (OuterVolumeSpecName: "kube-api-access-fwzxc") pod "f866fb32-3995-422b-a267-e0cc005d9c12" (UID: "f866fb32-3995-422b-a267-e0cc005d9c12"). InnerVolumeSpecName "kube-api-access-fwzxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.996564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdpz2\" (UniqueName: \"kubernetes.io/projected/e413986a-213e-41e2-ab3e-2d2e8899110a-kube-api-access-pdpz2\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:45 crc kubenswrapper[4810]: I1008 06:48:45.999631 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.011895 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f866fb32-3995-422b-a267-e0cc005d9c12" (UID: "f866fb32-3995-422b-a267-e0cc005d9c12"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.015944 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f866fb32-3995-422b-a267-e0cc005d9c12" (UID: "f866fb32-3995-422b-a267-e0cc005d9c12"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.027924 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-h7z72"] Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.034527 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.053111 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f866fb32-3995-422b-a267-e0cc005d9c12" (UID: "f866fb32-3995-422b-a267-e0cc005d9c12"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.056130 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-config" (OuterVolumeSpecName: "config") pod "f866fb32-3995-422b-a267-e0cc005d9c12" (UID: "f866fb32-3995-422b-a267-e0cc005d9c12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.057696 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f866fb32-3995-422b-a267-e0cc005d9c12" (UID: "f866fb32-3995-422b-a267-e0cc005d9c12"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.066397 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.066439 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwzxc\" (UniqueName: \"kubernetes.io/projected/f866fb32-3995-422b-a267-e0cc005d9c12-kube-api-access-fwzxc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.066456 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.066469 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.066480 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.066493 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f866fb32-3995-422b-a267-e0cc005d9c12-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.092463 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.112567 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.300240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerStarted","Data":"b5613fc4b756fcf92f58fee50d7c007c1e163f111c308da8ffefba505019b8f3"} Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.305043 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2x6lv" event={"ID":"0650b229-1dc2-454b-b2b5-62adc442b564","Type":"ContainerStarted","Data":"1496b7ed265837ae3927cb59497e3d39ba71ca3553ca631f821eef258bd3755f"} Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.305094 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2x6lv" event={"ID":"0650b229-1dc2-454b-b2b5-62adc442b564","Type":"ContainerStarted","Data":"7a7712de0350fce6c587d9ed992df0758f4a1be5acd6d7ae62d462eab486227b"} Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.315640 4810 generic.go:334] "Generic (PLEG): container finished" podID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerID="dc33a789cf086b565ca26150f8a066de6c4e57ed631aabf50ab45142b77970ef" exitCode=0 Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.315742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" event={"ID":"9d5fc753-8316-4e77-8e4a-c0ab288b6589","Type":"ContainerDied","Data":"dc33a789cf086b565ca26150f8a066de6c4e57ed631aabf50ab45142b77970ef"} Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.315787 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" event={"ID":"9d5fc753-8316-4e77-8e4a-c0ab288b6589","Type":"ContainerStarted","Data":"0f50f021b43579d636062d4ec7e244b3fcf631eb1d2bf79db6e4ec97b62f2eca"} Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.320471 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795846498c-pp6hm" event={"ID":"f866fb32-3995-422b-a267-e0cc005d9c12","Type":"ContainerDied","Data":"8f023d7ba39594a3f04d18e9eb23dee8c24c749920b09c1301d09ab133ad5342"} Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.320520 4810 scope.go:117] "RemoveContainer" containerID="6faccb5e7c8065d98bfa6fafd38dd3b7eed39278107ded21bca68560f68b9cc3" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.320587 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795846498c-pp6hm" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.324783 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b4bfdd7f7-9m56r" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.324916 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h7z72" event={"ID":"22742881-9836-4dbb-bd8c-deba4194d7c3","Type":"ContainerStarted","Data":"0a126c6c7d59bed5a5cfee3aeead62f6837383badf81572632ead27ed551843c"} Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.346123 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2x6lv" podStartSLOduration=2.346097959 podStartE2EDuration="2.346097959s" podCreationTimestamp="2025-10-08 06:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:46.338420116 +0000 UTC m=+1028.972859856" watchObservedRunningTime="2025-10-08 06:48:46.346097959 +0000 UTC m=+1028.980537729" Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.444925 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795846498c-pp6hm"] Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.493958 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-795846498c-pp6hm"] Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.542757 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b4bfdd7f7-9m56r"] Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.554017 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b4bfdd7f7-9m56r"] Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.684332 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:46 crc kubenswrapper[4810]: W1008 06:48:46.695951 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f7195c4_3f0e_412c_bf83_2862e38a22d5.slice/crio-a9dbc9ce500d8b761fb7ac32b17fd38597813c36603eec0599898937f1bb36e0 WatchSource:0}: Error finding container a9dbc9ce500d8b761fb7ac32b17fd38597813c36603eec0599898937f1bb36e0: Status 404 returned error can't find the container with id a9dbc9ce500d8b761fb7ac32b17fd38597813c36603eec0599898937f1bb36e0 Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.820388 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:46 crc kubenswrapper[4810]: I1008 06:48:46.993851 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:47 crc kubenswrapper[4810]: I1008 06:48:47.035301 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:48:47 crc kubenswrapper[4810]: I1008 06:48:47.082149 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:47 crc kubenswrapper[4810]: I1008 06:48:47.342747 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e413986a-213e-41e2-ab3e-2d2e8899110a","Type":"ContainerStarted","Data":"ff2061bc0793793a3c6f2a6cfd446efa2392f356d52a50bc04c834c4008a47fe"} Oct 08 06:48:47 crc kubenswrapper[4810]: I1008 06:48:47.353762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" event={"ID":"9d5fc753-8316-4e77-8e4a-c0ab288b6589","Type":"ContainerStarted","Data":"cafd24014584c1d4c9cf578ffee08a99f18df72ab6bfd6c4794466651cfa79fe"} Oct 08 06:48:47 crc kubenswrapper[4810]: I1008 06:48:47.353912 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:47 crc kubenswrapper[4810]: I1008 06:48:47.355614 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6f7195c4-3f0e-412c-bf83-2862e38a22d5","Type":"ContainerStarted","Data":"a9dbc9ce500d8b761fb7ac32b17fd38597813c36603eec0599898937f1bb36e0"} Oct 08 06:48:47 crc kubenswrapper[4810]: I1008 06:48:47.381137 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" podStartSLOduration=3.381118266 podStartE2EDuration="3.381118266s" podCreationTimestamp="2025-10-08 06:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:47.377265929 +0000 UTC m=+1030.011705689" watchObservedRunningTime="2025-10-08 06:48:47.381118266 +0000 UTC m=+1030.015558006" Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.095458 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f48bbd0-a1c0-49ae-a133-d1c78b4869f9" path="/var/lib/kubelet/pods/6f48bbd0-a1c0-49ae-a133-d1c78b4869f9/volumes" Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.096594 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f866fb32-3995-422b-a267-e0cc005d9c12" path="/var/lib/kubelet/pods/f866fb32-3995-422b-a267-e0cc005d9c12/volumes" Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.373780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6f7195c4-3f0e-412c-bf83-2862e38a22d5","Type":"ContainerStarted","Data":"3fdd681e4915b8c2b3ef38732d91f226c01eb002a8b494a8d4c8cf453e8db459"} Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.373828 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6f7195c4-3f0e-412c-bf83-2862e38a22d5","Type":"ContainerStarted","Data":"42985e588039e90c5c6ed14a79d1e3750175c7b70501a1437777e6d99251b196"} Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.373939 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-log" containerID="cri-o://42985e588039e90c5c6ed14a79d1e3750175c7b70501a1437777e6d99251b196" gracePeriod=30 Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.374094 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-httpd" containerID="cri-o://3fdd681e4915b8c2b3ef38732d91f226c01eb002a8b494a8d4c8cf453e8db459" gracePeriod=30 Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.378682 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e413986a-213e-41e2-ab3e-2d2e8899110a","Type":"ContainerStarted","Data":"3c624b43a6c9892e17ecd9221fb2d6d0de82c4169601068172ec1e95d88ee2a5"} Oct 08 06:48:48 crc kubenswrapper[4810]: I1008 06:48:48.408371 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.408344867 podStartE2EDuration="4.408344867s" podCreationTimestamp="2025-10-08 06:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:48.399628855 +0000 UTC m=+1031.034068615" watchObservedRunningTime="2025-10-08 06:48:48.408344867 +0000 UTC m=+1031.042784617" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.005529 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1485-account-create-687t8"] Oct 08 06:48:49 crc kubenswrapper[4810]: E1008 06:48:49.005878 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f866fb32-3995-422b-a267-e0cc005d9c12" containerName="init" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.005891 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f866fb32-3995-422b-a267-e0cc005d9c12" containerName="init" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.006096 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f866fb32-3995-422b-a267-e0cc005d9c12" containerName="init" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.006775 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1485-account-create-687t8" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.022625 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1485-account-create-687t8"] Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.026788 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.152741 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9s2n\" (UniqueName: \"kubernetes.io/projected/1d65fbac-67f5-445f-9fd6-56ae7828fba8-kube-api-access-j9s2n\") pod \"barbican-1485-account-create-687t8\" (UID: \"1d65fbac-67f5-445f-9fd6-56ae7828fba8\") " pod="openstack/barbican-1485-account-create-687t8" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.195810 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-95b0-account-create-txvcv"] Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.197045 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-95b0-account-create-txvcv" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.202089 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.209246 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-95b0-account-create-txvcv"] Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.255631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5j8q\" (UniqueName: \"kubernetes.io/projected/0babfa70-91ca-45ad-afb6-aa40d2300144-kube-api-access-t5j8q\") pod \"cinder-95b0-account-create-txvcv\" (UID: \"0babfa70-91ca-45ad-afb6-aa40d2300144\") " pod="openstack/cinder-95b0-account-create-txvcv" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.255730 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9s2n\" (UniqueName: \"kubernetes.io/projected/1d65fbac-67f5-445f-9fd6-56ae7828fba8-kube-api-access-j9s2n\") pod \"barbican-1485-account-create-687t8\" (UID: \"1d65fbac-67f5-445f-9fd6-56ae7828fba8\") " pod="openstack/barbican-1485-account-create-687t8" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.283025 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9s2n\" (UniqueName: \"kubernetes.io/projected/1d65fbac-67f5-445f-9fd6-56ae7828fba8-kube-api-access-j9s2n\") pod \"barbican-1485-account-create-687t8\" (UID: \"1d65fbac-67f5-445f-9fd6-56ae7828fba8\") " pod="openstack/barbican-1485-account-create-687t8" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.332802 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1485-account-create-687t8" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.359269 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5j8q\" (UniqueName: \"kubernetes.io/projected/0babfa70-91ca-45ad-afb6-aa40d2300144-kube-api-access-t5j8q\") pod \"cinder-95b0-account-create-txvcv\" (UID: \"0babfa70-91ca-45ad-afb6-aa40d2300144\") " pod="openstack/cinder-95b0-account-create-txvcv" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.391167 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5j8q\" (UniqueName: \"kubernetes.io/projected/0babfa70-91ca-45ad-afb6-aa40d2300144-kube-api-access-t5j8q\") pod \"cinder-95b0-account-create-txvcv\" (UID: \"0babfa70-91ca-45ad-afb6-aa40d2300144\") " pod="openstack/cinder-95b0-account-create-txvcv" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.416291 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b73c-account-create-6nkpk"] Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.418062 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b73c-account-create-6nkpk" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.421719 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.424534 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-log" containerID="cri-o://3c624b43a6c9892e17ecd9221fb2d6d0de82c4169601068172ec1e95d88ee2a5" gracePeriod=30 Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.424674 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e413986a-213e-41e2-ab3e-2d2e8899110a","Type":"ContainerStarted","Data":"4bdd726c2580410153c0f4a2c209496de0518a15b8fecb05b3aa11e4ba9d0091"} Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.424775 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-httpd" containerID="cri-o://4bdd726c2580410153c0f4a2c209496de0518a15b8fecb05b3aa11e4ba9d0091" gracePeriod=30 Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.429224 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b73c-account-create-6nkpk"] Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.441920 4810 generic.go:334] "Generic (PLEG): container finished" podID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerID="3fdd681e4915b8c2b3ef38732d91f226c01eb002a8b494a8d4c8cf453e8db459" exitCode=143 Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.441991 4810 generic.go:334] "Generic (PLEG): container finished" podID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerID="42985e588039e90c5c6ed14a79d1e3750175c7b70501a1437777e6d99251b196" exitCode=143 Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.442024 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6f7195c4-3f0e-412c-bf83-2862e38a22d5","Type":"ContainerDied","Data":"3fdd681e4915b8c2b3ef38732d91f226c01eb002a8b494a8d4c8cf453e8db459"} Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.442104 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6f7195c4-3f0e-412c-bf83-2862e38a22d5","Type":"ContainerDied","Data":"42985e588039e90c5c6ed14a79d1e3750175c7b70501a1437777e6d99251b196"} Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.527669 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-95b0-account-create-txvcv" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.563056 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c97x\" (UniqueName: \"kubernetes.io/projected/a6363724-0c41-4b54-89e6-c672b91bbe98-kube-api-access-4c97x\") pod \"neutron-b73c-account-create-6nkpk\" (UID: \"a6363724-0c41-4b54-89e6-c672b91bbe98\") " pod="openstack/neutron-b73c-account-create-6nkpk" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.665834 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c97x\" (UniqueName: \"kubernetes.io/projected/a6363724-0c41-4b54-89e6-c672b91bbe98-kube-api-access-4c97x\") pod \"neutron-b73c-account-create-6nkpk\" (UID: \"a6363724-0c41-4b54-89e6-c672b91bbe98\") " pod="openstack/neutron-b73c-account-create-6nkpk" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.692537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c97x\" (UniqueName: \"kubernetes.io/projected/a6363724-0c41-4b54-89e6-c672b91bbe98-kube-api-access-4c97x\") pod \"neutron-b73c-account-create-6nkpk\" (UID: \"a6363724-0c41-4b54-89e6-c672b91bbe98\") " pod="openstack/neutron-b73c-account-create-6nkpk" Oct 08 06:48:49 crc kubenswrapper[4810]: I1008 06:48:49.752092 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b73c-account-create-6nkpk" Oct 08 06:48:50 crc kubenswrapper[4810]: I1008 06:48:50.465820 4810 generic.go:334] "Generic (PLEG): container finished" podID="0650b229-1dc2-454b-b2b5-62adc442b564" containerID="1496b7ed265837ae3927cb59497e3d39ba71ca3553ca631f821eef258bd3755f" exitCode=0 Oct 08 06:48:50 crc kubenswrapper[4810]: I1008 06:48:50.465999 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2x6lv" event={"ID":"0650b229-1dc2-454b-b2b5-62adc442b564","Type":"ContainerDied","Data":"1496b7ed265837ae3927cb59497e3d39ba71ca3553ca631f821eef258bd3755f"} Oct 08 06:48:50 crc kubenswrapper[4810]: I1008 06:48:50.473837 4810 generic.go:334] "Generic (PLEG): container finished" podID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerID="4bdd726c2580410153c0f4a2c209496de0518a15b8fecb05b3aa11e4ba9d0091" exitCode=0 Oct 08 06:48:50 crc kubenswrapper[4810]: I1008 06:48:50.473869 4810 generic.go:334] "Generic (PLEG): container finished" podID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerID="3c624b43a6c9892e17ecd9221fb2d6d0de82c4169601068172ec1e95d88ee2a5" exitCode=143 Oct 08 06:48:50 crc kubenswrapper[4810]: I1008 06:48:50.473902 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e413986a-213e-41e2-ab3e-2d2e8899110a","Type":"ContainerDied","Data":"4bdd726c2580410153c0f4a2c209496de0518a15b8fecb05b3aa11e4ba9d0091"} Oct 08 06:48:50 crc kubenswrapper[4810]: I1008 06:48:50.473942 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e413986a-213e-41e2-ab3e-2d2e8899110a","Type":"ContainerDied","Data":"3c624b43a6c9892e17ecd9221fb2d6d0de82c4169601068172ec1e95d88ee2a5"} Oct 08 06:48:50 crc kubenswrapper[4810]: I1008 06:48:50.501776 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.50175646 podStartE2EDuration="6.50175646s" podCreationTimestamp="2025-10-08 06:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:49.460313655 +0000 UTC m=+1032.094753395" watchObservedRunningTime="2025-10-08 06:48:50.50175646 +0000 UTC m=+1033.136196200" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.741857 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.742954 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.742502 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.867915 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-config-data\") pod \"e413986a-213e-41e2-ab3e-2d2e8899110a\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868466 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-combined-ca-bundle\") pod \"e413986a-213e-41e2-ab3e-2d2e8899110a\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868533 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-logs\") pod \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868569 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"e413986a-213e-41e2-ab3e-2d2e8899110a\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868598 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-httpd-run\") pod \"e413986a-213e-41e2-ab3e-2d2e8899110a\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868651 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-combined-ca-bundle\") pod \"0650b229-1dc2-454b-b2b5-62adc442b564\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868683 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-credential-keys\") pod \"0650b229-1dc2-454b-b2b5-62adc442b564\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868735 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpg5b\" (UniqueName: \"kubernetes.io/projected/0650b229-1dc2-454b-b2b5-62adc442b564-kube-api-access-cpg5b\") pod \"0650b229-1dc2-454b-b2b5-62adc442b564\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868756 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6st2w\" (UniqueName: \"kubernetes.io/projected/6f7195c4-3f0e-412c-bf83-2862e38a22d5-kube-api-access-6st2w\") pod \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868787 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-scripts\") pod \"0650b229-1dc2-454b-b2b5-62adc442b564\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868805 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-httpd-run\") pod \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-scripts\") pod \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868857 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-fernet-keys\") pod \"0650b229-1dc2-454b-b2b5-62adc442b564\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868880 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-config-data\") pod \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868924 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-combined-ca-bundle\") pod \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\" (UID: \"6f7195c4-3f0e-412c-bf83-2862e38a22d5\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.868953 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-scripts\") pod \"e413986a-213e-41e2-ab3e-2d2e8899110a\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.869013 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdpz2\" (UniqueName: \"kubernetes.io/projected/e413986a-213e-41e2-ab3e-2d2e8899110a-kube-api-access-pdpz2\") pod \"e413986a-213e-41e2-ab3e-2d2e8899110a\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.869049 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-logs\") pod \"e413986a-213e-41e2-ab3e-2d2e8899110a\" (UID: \"e413986a-213e-41e2-ab3e-2d2e8899110a\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.869083 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-config-data\") pod \"0650b229-1dc2-454b-b2b5-62adc442b564\" (UID: \"0650b229-1dc2-454b-b2b5-62adc442b564\") " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.869169 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-logs" (OuterVolumeSpecName: "logs") pod "6f7195c4-3f0e-412c-bf83-2862e38a22d5" (UID: "6f7195c4-3f0e-412c-bf83-2862e38a22d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.869176 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e413986a-213e-41e2-ab3e-2d2e8899110a" (UID: "e413986a-213e-41e2-ab3e-2d2e8899110a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.869650 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.869667 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.870996 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6f7195c4-3f0e-412c-bf83-2862e38a22d5" (UID: "6f7195c4-3f0e-412c-bf83-2862e38a22d5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.885358 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "e413986a-213e-41e2-ab3e-2d2e8899110a" (UID: "e413986a-213e-41e2-ab3e-2d2e8899110a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.895716 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-logs" (OuterVolumeSpecName: "logs") pod "e413986a-213e-41e2-ab3e-2d2e8899110a" (UID: "e413986a-213e-41e2-ab3e-2d2e8899110a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.897522 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "6f7195c4-3f0e-412c-bf83-2862e38a22d5" (UID: "6f7195c4-3f0e-412c-bf83-2862e38a22d5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.925015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-scripts" (OuterVolumeSpecName: "scripts") pod "0650b229-1dc2-454b-b2b5-62adc442b564" (UID: "0650b229-1dc2-454b-b2b5-62adc442b564"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.944450 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0650b229-1dc2-454b-b2b5-62adc442b564-kube-api-access-cpg5b" (OuterVolumeSpecName: "kube-api-access-cpg5b") pod "0650b229-1dc2-454b-b2b5-62adc442b564" (UID: "0650b229-1dc2-454b-b2b5-62adc442b564"). InnerVolumeSpecName "kube-api-access-cpg5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.944593 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0650b229-1dc2-454b-b2b5-62adc442b564" (UID: "0650b229-1dc2-454b-b2b5-62adc442b564"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.946014 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-scripts" (OuterVolumeSpecName: "scripts") pod "e413986a-213e-41e2-ab3e-2d2e8899110a" (UID: "e413986a-213e-41e2-ab3e-2d2e8899110a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.949844 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0650b229-1dc2-454b-b2b5-62adc442b564" (UID: "0650b229-1dc2-454b-b2b5-62adc442b564"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.949869 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-scripts" (OuterVolumeSpecName: "scripts") pod "6f7195c4-3f0e-412c-bf83-2862e38a22d5" (UID: "6f7195c4-3f0e-412c-bf83-2862e38a22d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.950203 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f7195c4-3f0e-412c-bf83-2862e38a22d5-kube-api-access-6st2w" (OuterVolumeSpecName: "kube-api-access-6st2w") pod "6f7195c4-3f0e-412c-bf83-2862e38a22d5" (UID: "6f7195c4-3f0e-412c-bf83-2862e38a22d5"). InnerVolumeSpecName "kube-api-access-6st2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.957749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e413986a-213e-41e2-ab3e-2d2e8899110a-kube-api-access-pdpz2" (OuterVolumeSpecName: "kube-api-access-pdpz2") pod "e413986a-213e-41e2-ab3e-2d2e8899110a" (UID: "e413986a-213e-41e2-ab3e-2d2e8899110a"). InnerVolumeSpecName "kube-api-access-pdpz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972130 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972170 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdpz2\" (UniqueName: \"kubernetes.io/projected/e413986a-213e-41e2-ab3e-2d2e8899110a-kube-api-access-pdpz2\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972183 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e413986a-213e-41e2-ab3e-2d2e8899110a-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972204 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972217 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972226 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972236 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpg5b\" (UniqueName: \"kubernetes.io/projected/0650b229-1dc2-454b-b2b5-62adc442b564-kube-api-access-cpg5b\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972244 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6st2w\" (UniqueName: \"kubernetes.io/projected/6f7195c4-3f0e-412c-bf83-2862e38a22d5-kube-api-access-6st2w\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972252 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972260 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6f7195c4-3f0e-412c-bf83-2862e38a22d5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972267 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.972275 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:53 crc kubenswrapper[4810]: I1008 06:48:53.994810 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-config-data" (OuterVolumeSpecName: "config-data") pod "0650b229-1dc2-454b-b2b5-62adc442b564" (UID: "0650b229-1dc2-454b-b2b5-62adc442b564"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.034185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e413986a-213e-41e2-ab3e-2d2e8899110a" (UID: "e413986a-213e-41e2-ab3e-2d2e8899110a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.039182 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.045800 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.068126 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f7195c4-3f0e-412c-bf83-2862e38a22d5" (UID: "6f7195c4-3f0e-412c-bf83-2862e38a22d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.082423 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.082454 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.082465 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.082477 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.082487 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.132183 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0650b229-1dc2-454b-b2b5-62adc442b564" (UID: "0650b229-1dc2-454b-b2b5-62adc442b564"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.168307 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-config-data" (OuterVolumeSpecName: "config-data") pod "6f7195c4-3f0e-412c-bf83-2862e38a22d5" (UID: "6f7195c4-3f0e-412c-bf83-2862e38a22d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.170460 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-config-data" (OuterVolumeSpecName: "config-data") pod "e413986a-213e-41e2-ab3e-2d2e8899110a" (UID: "e413986a-213e-41e2-ab3e-2d2e8899110a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.184721 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0650b229-1dc2-454b-b2b5-62adc442b564-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.184761 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f7195c4-3f0e-412c-bf83-2862e38a22d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.184772 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413986a-213e-41e2-ab3e-2d2e8899110a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:54 crc kubenswrapper[4810]: E1008 06:48:54.405972 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0650b229_1dc2_454b_b2b5_62adc442b564.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0650b229_1dc2_454b_b2b5_62adc442b564.slice/crio-7a7712de0350fce6c587d9ed992df0758f4a1be5acd6d7ae62d462eab486227b\": RecentStats: unable to find data in memory cache]" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.505007 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1485-account-create-687t8"] Oct 08 06:48:54 crc kubenswrapper[4810]: W1008 06:48:54.507851 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d65fbac_67f5_445f_9fd6_56ae7828fba8.slice/crio-a3764bcaaa54c8ab7f76147a1eff870d8993a3530ab9f781a1da4d9e3fb7a899 WatchSource:0}: Error finding container a3764bcaaa54c8ab7f76147a1eff870d8993a3530ab9f781a1da4d9e3fb7a899: Status 404 returned error can't find the container with id a3764bcaaa54c8ab7f76147a1eff870d8993a3530ab9f781a1da4d9e3fb7a899 Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.528393 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e413986a-213e-41e2-ab3e-2d2e8899110a","Type":"ContainerDied","Data":"ff2061bc0793793a3c6f2a6cfd446efa2392f356d52a50bc04c834c4008a47fe"} Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.528460 4810 scope.go:117] "RemoveContainer" containerID="4bdd726c2580410153c0f4a2c209496de0518a15b8fecb05b3aa11e4ba9d0091" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.528663 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.539455 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h7z72" event={"ID":"22742881-9836-4dbb-bd8c-deba4194d7c3","Type":"ContainerStarted","Data":"76a5e6948d4daf32096a3e2e398fd4087c887c28e01b46a3346cda1ecc593c0c"} Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.547050 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerStarted","Data":"8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66"} Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.556863 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2x6lv" event={"ID":"0650b229-1dc2-454b-b2b5-62adc442b564","Type":"ContainerDied","Data":"7a7712de0350fce6c587d9ed992df0758f4a1be5acd6d7ae62d462eab486227b"} Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.556920 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a7712de0350fce6c587d9ed992df0758f4a1be5acd6d7ae62d462eab486227b" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.557028 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2x6lv" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.560340 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6f7195c4-3f0e-412c-bf83-2862e38a22d5","Type":"ContainerDied","Data":"a9dbc9ce500d8b761fb7ac32b17fd38597813c36603eec0599898937f1bb36e0"} Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.560504 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.567087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1485-account-create-687t8" event={"ID":"1d65fbac-67f5-445f-9fd6-56ae7828fba8","Type":"ContainerStarted","Data":"a3764bcaaa54c8ab7f76147a1eff870d8993a3530ab9f781a1da4d9e3fb7a899"} Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.578382 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-h7z72" podStartSLOduration=2.9982480540000003 podStartE2EDuration="10.578364649s" podCreationTimestamp="2025-10-08 06:48:44 +0000 UTC" firstStartedPulling="2025-10-08 06:48:46.041095677 +0000 UTC m=+1028.675535417" lastFinishedPulling="2025-10-08 06:48:53.621212272 +0000 UTC m=+1036.255652012" observedRunningTime="2025-10-08 06:48:54.577887055 +0000 UTC m=+1037.212326795" watchObservedRunningTime="2025-10-08 06:48:54.578364649 +0000 UTC m=+1037.212804389" Oct 08 06:48:54 crc kubenswrapper[4810]: W1008 06:48:54.596165 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6363724_0c41_4b54_89e6_c672b91bbe98.slice/crio-6b8162095ccf49350ff0d4c931f3621aa617a20c0b5493195de3ba1f78f848cb WatchSource:0}: Error finding container 6b8162095ccf49350ff0d4c931f3621aa617a20c0b5493195de3ba1f78f848cb: Status 404 returned error can't find the container with id 6b8162095ccf49350ff0d4c931f3621aa617a20c0b5493195de3ba1f78f848cb Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.620068 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b73c-account-create-6nkpk"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.629831 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-95b0-account-create-txvcv"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.726248 4810 scope.go:117] "RemoveContainer" containerID="3c624b43a6c9892e17ecd9221fb2d6d0de82c4169601068172ec1e95d88ee2a5" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.767366 4810 scope.go:117] "RemoveContainer" containerID="3fdd681e4915b8c2b3ef38732d91f226c01eb002a8b494a8d4c8cf453e8db459" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.797896 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.837423 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.887624 4810 scope.go:117] "RemoveContainer" containerID="42985e588039e90c5c6ed14a79d1e3750175c7b70501a1437777e6d99251b196" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.910766 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.931615 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.945680 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: E1008 06:48:54.949800 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-httpd" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.949829 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-httpd" Oct 08 06:48:54 crc kubenswrapper[4810]: E1008 06:48:54.949840 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-httpd" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.949848 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-httpd" Oct 08 06:48:54 crc kubenswrapper[4810]: E1008 06:48:54.949868 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-log" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.949875 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-log" Oct 08 06:48:54 crc kubenswrapper[4810]: E1008 06:48:54.949911 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0650b229-1dc2-454b-b2b5-62adc442b564" containerName="keystone-bootstrap" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.949918 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0650b229-1dc2-454b-b2b5-62adc442b564" containerName="keystone-bootstrap" Oct 08 06:48:54 crc kubenswrapper[4810]: E1008 06:48:54.949926 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-log" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.949932 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-log" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.950166 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-log" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.950187 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" containerName="glance-httpd" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.950198 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-log" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.950207 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0650b229-1dc2-454b-b2b5-62adc442b564" containerName="keystone-bootstrap" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.950222 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" containerName="glance-httpd" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.951397 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.953329 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.957533 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.958578 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.958799 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.958914 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.959172 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6dmpd" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.962212 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.967868 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.981331 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:54 crc kubenswrapper[4810]: I1008 06:48:54.993093 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2x6lv"] Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.003389 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.011077 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2x6lv"] Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.020515 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-m2q9f"] Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.022317 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.030009 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m2q9f"] Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.037595 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.037999 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tb5ft" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.038188 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.038393 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123198 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123346 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123366 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-fernet-keys\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123398 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123430 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kprpx\" (UniqueName: \"kubernetes.io/projected/5632ea60-f549-438f-a244-50f811fefc1a-kube-api-access-kprpx\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123467 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123489 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123505 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-credential-keys\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123528 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123547 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-combined-ca-bundle\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123565 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ktf2\" (UniqueName: \"kubernetes.io/projected/9d3b9863-3ccd-4bda-93da-ee17ac885040-kube-api-access-2ktf2\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123583 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb4x2\" (UniqueName: \"kubernetes.io/projected/5ad9084b-295c-49da-a76c-aa3cde9ffa29-kube-api-access-lb4x2\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123598 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-scripts\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123614 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123684 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-scripts\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-config-data\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123744 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.123771 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-logs\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.124643 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-logs\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.124798 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.212853 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-564965cbfc-6q8bc"] Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.213109 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" podUID="78b447d3-3cd5-440a-8879-02c91e23e043" containerName="dnsmasq-dns" containerID="cri-o://b7c833b41867e5b68178094e9963af557ce11cc0418b0f74d99fcdb749a941cc" gracePeriod=10 Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226365 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-config-data\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226417 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226441 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-logs\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-logs\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226533 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226584 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226601 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-fernet-keys\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226632 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226684 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kprpx\" (UniqueName: \"kubernetes.io/projected/5632ea60-f549-438f-a244-50f811fefc1a-kube-api-access-kprpx\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226715 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226732 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226767 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-credential-keys\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226788 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-combined-ca-bundle\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ktf2\" (UniqueName: \"kubernetes.io/projected/9d3b9863-3ccd-4bda-93da-ee17ac885040-kube-api-access-2ktf2\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226860 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb4x2\" (UniqueName: \"kubernetes.io/projected/5ad9084b-295c-49da-a76c-aa3cde9ffa29-kube-api-access-lb4x2\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226874 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-scripts\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.226924 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.227016 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-scripts\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.228339 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.228899 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.230092 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.232981 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-logs\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.233291 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-logs\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.234697 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.245492 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-fernet-keys\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.246567 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.249191 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-scripts\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.260835 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-credential-keys\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.261327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.261403 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-config-data\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.261769 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.261788 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.261927 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ktf2\" (UniqueName: \"kubernetes.io/projected/9d3b9863-3ccd-4bda-93da-ee17ac885040-kube-api-access-2ktf2\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.261997 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-combined-ca-bundle\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.262048 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-scripts\") pod \"keystone-bootstrap-m2q9f\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.262356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.265017 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.266048 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.272376 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb4x2\" (UniqueName: \"kubernetes.io/projected/5ad9084b-295c-49da-a76c-aa3cde9ffa29-kube-api-access-lb4x2\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.272436 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kprpx\" (UniqueName: \"kubernetes.io/projected/5632ea60-f549-438f-a244-50f811fefc1a-kube-api-access-kprpx\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.291151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.296203 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.320791 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.348164 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.599932 4810 generic.go:334] "Generic (PLEG): container finished" podID="78b447d3-3cd5-440a-8879-02c91e23e043" containerID="b7c833b41867e5b68178094e9963af557ce11cc0418b0f74d99fcdb749a941cc" exitCode=0 Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.600264 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" event={"ID":"78b447d3-3cd5-440a-8879-02c91e23e043","Type":"ContainerDied","Data":"b7c833b41867e5b68178094e9963af557ce11cc0418b0f74d99fcdb749a941cc"} Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.601447 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.602216 4810 generic.go:334] "Generic (PLEG): container finished" podID="0babfa70-91ca-45ad-afb6-aa40d2300144" containerID="6a3956be0e123bf75016806de880277e6c672f2d8443f2f7b9d2fbda532d3217" exitCode=0 Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.602334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-95b0-account-create-txvcv" event={"ID":"0babfa70-91ca-45ad-afb6-aa40d2300144","Type":"ContainerDied","Data":"6a3956be0e123bf75016806de880277e6c672f2d8443f2f7b9d2fbda532d3217"} Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.602425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-95b0-account-create-txvcv" event={"ID":"0babfa70-91ca-45ad-afb6-aa40d2300144","Type":"ContainerStarted","Data":"160a09cb18861d3fc5c7e43e8458d2ffae624dfe4c9a58b4d7e0b579193d3148"} Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.604196 4810 generic.go:334] "Generic (PLEG): container finished" podID="1d65fbac-67f5-445f-9fd6-56ae7828fba8" containerID="541ed13448c8e2c968839050706935ecc2bebbb1bbc55d3025a7f620c1798afd" exitCode=0 Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.604353 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1485-account-create-687t8" event={"ID":"1d65fbac-67f5-445f-9fd6-56ae7828fba8","Type":"ContainerDied","Data":"541ed13448c8e2c968839050706935ecc2bebbb1bbc55d3025a7f620c1798afd"} Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.613510 4810 generic.go:334] "Generic (PLEG): container finished" podID="a6363724-0c41-4b54-89e6-c672b91bbe98" containerID="8b1378e928905e98ed227c844df3846d51fa404c6a1cb6f7629ca5ddef749cff" exitCode=0 Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.613576 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b73c-account-create-6nkpk" event={"ID":"a6363724-0c41-4b54-89e6-c672b91bbe98","Type":"ContainerDied","Data":"8b1378e928905e98ed227c844df3846d51fa404c6a1cb6f7629ca5ddef749cff"} Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.613650 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b73c-account-create-6nkpk" event={"ID":"a6363724-0c41-4b54-89e6-c672b91bbe98","Type":"ContainerStarted","Data":"6b8162095ccf49350ff0d4c931f3621aa617a20c0b5493195de3ba1f78f848cb"} Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.805154 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.940731 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-sb\") pod \"78b447d3-3cd5-440a-8879-02c91e23e043\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.940780 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-nb\") pod \"78b447d3-3cd5-440a-8879-02c91e23e043\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.940843 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-swift-storage-0\") pod \"78b447d3-3cd5-440a-8879-02c91e23e043\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.940880 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nc6k\" (UniqueName: \"kubernetes.io/projected/78b447d3-3cd5-440a-8879-02c91e23e043-kube-api-access-8nc6k\") pod \"78b447d3-3cd5-440a-8879-02c91e23e043\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.941008 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-config\") pod \"78b447d3-3cd5-440a-8879-02c91e23e043\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.941033 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-svc\") pod \"78b447d3-3cd5-440a-8879-02c91e23e043\" (UID: \"78b447d3-3cd5-440a-8879-02c91e23e043\") " Oct 08 06:48:55 crc kubenswrapper[4810]: I1008 06:48:55.950181 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78b447d3-3cd5-440a-8879-02c91e23e043-kube-api-access-8nc6k" (OuterVolumeSpecName: "kube-api-access-8nc6k") pod "78b447d3-3cd5-440a-8879-02c91e23e043" (UID: "78b447d3-3cd5-440a-8879-02c91e23e043"). InnerVolumeSpecName "kube-api-access-8nc6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.023010 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78b447d3-3cd5-440a-8879-02c91e23e043" (UID: "78b447d3-3cd5-440a-8879-02c91e23e043"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.028662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "78b447d3-3cd5-440a-8879-02c91e23e043" (UID: "78b447d3-3cd5-440a-8879-02c91e23e043"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.030479 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "78b447d3-3cd5-440a-8879-02c91e23e043" (UID: "78b447d3-3cd5-440a-8879-02c91e23e043"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.033952 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m2q9f"] Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.035547 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "78b447d3-3cd5-440a-8879-02c91e23e043" (UID: "78b447d3-3cd5-440a-8879-02c91e23e043"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.046080 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.046126 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.046142 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.046157 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nc6k\" (UniqueName: \"kubernetes.io/projected/78b447d3-3cd5-440a-8879-02c91e23e043-kube-api-access-8nc6k\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.046173 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.053609 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-config" (OuterVolumeSpecName: "config") pod "78b447d3-3cd5-440a-8879-02c91e23e043" (UID: "78b447d3-3cd5-440a-8879-02c91e23e043"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.090790 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0650b229-1dc2-454b-b2b5-62adc442b564" path="/var/lib/kubelet/pods/0650b229-1dc2-454b-b2b5-62adc442b564/volumes" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.091480 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f7195c4-3f0e-412c-bf83-2862e38a22d5" path="/var/lib/kubelet/pods/6f7195c4-3f0e-412c-bf83-2862e38a22d5/volumes" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.092421 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e413986a-213e-41e2-ab3e-2d2e8899110a" path="/var/lib/kubelet/pods/e413986a-213e-41e2-ab3e-2d2e8899110a/volumes" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.150349 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b447d3-3cd5-440a-8879-02c91e23e043-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.267126 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.357255 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:48:56 crc kubenswrapper[4810]: W1008 06:48:56.394099 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d3b9863_3ccd_4bda_93da_ee17ac885040.slice/crio-8744b5ebe7a4650dca24b3baf7106f29051ba778b3d32ce3ffed70d395e9a149 WatchSource:0}: Error finding container 8744b5ebe7a4650dca24b3baf7106f29051ba778b3d32ce3ffed70d395e9a149: Status 404 returned error can't find the container with id 8744b5ebe7a4650dca24b3baf7106f29051ba778b3d32ce3ffed70d395e9a149 Oct 08 06:48:56 crc kubenswrapper[4810]: W1008 06:48:56.395935 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ad9084b_295c_49da_a76c_aa3cde9ffa29.slice/crio-b75e1d5216e2d3a26d5eae25932d337b3197c0f0f2868a6f3149cc4b7eee522d WatchSource:0}: Error finding container b75e1d5216e2d3a26d5eae25932d337b3197c0f0f2868a6f3149cc4b7eee522d: Status 404 returned error can't find the container with id b75e1d5216e2d3a26d5eae25932d337b3197c0f0f2868a6f3149cc4b7eee522d Oct 08 06:48:56 crc kubenswrapper[4810]: W1008 06:48:56.397845 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5632ea60_f549_438f_a244_50f811fefc1a.slice/crio-e994bf63e6e41b4c7a28ffa2e35ce5fe6d57d7da24822021104f2e58454e1eb6 WatchSource:0}: Error finding container e994bf63e6e41b4c7a28ffa2e35ce5fe6d57d7da24822021104f2e58454e1eb6: Status 404 returned error can't find the container with id e994bf63e6e41b4c7a28ffa2e35ce5fe6d57d7da24822021104f2e58454e1eb6 Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.696334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5632ea60-f549-438f-a244-50f811fefc1a","Type":"ContainerStarted","Data":"e994bf63e6e41b4c7a28ffa2e35ce5fe6d57d7da24822021104f2e58454e1eb6"} Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.700883 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ad9084b-295c-49da-a76c-aa3cde9ffa29","Type":"ContainerStarted","Data":"b75e1d5216e2d3a26d5eae25932d337b3197c0f0f2868a6f3149cc4b7eee522d"} Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.727422 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.728446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564965cbfc-6q8bc" event={"ID":"78b447d3-3cd5-440a-8879-02c91e23e043","Type":"ContainerDied","Data":"9c13e6604196f164be6eb62fd044882b7ac444369f7f4a64bb37fceb06ca79cb"} Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.728498 4810 scope.go:117] "RemoveContainer" containerID="b7c833b41867e5b68178094e9963af557ce11cc0418b0f74d99fcdb749a941cc" Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.743844 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2q9f" event={"ID":"9d3b9863-3ccd-4bda-93da-ee17ac885040","Type":"ContainerStarted","Data":"8744b5ebe7a4650dca24b3baf7106f29051ba778b3d32ce3ffed70d395e9a149"} Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.784024 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-564965cbfc-6q8bc"] Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.796385 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-564965cbfc-6q8bc"] Oct 08 06:48:56 crc kubenswrapper[4810]: I1008 06:48:56.837135 4810 scope.go:117] "RemoveContainer" containerID="6cd4b9c52d92efd669812c350f816e6f8d083bdac9339adf49e3105f6eb37d82" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.149417 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b73c-account-create-6nkpk" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.311698 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c97x\" (UniqueName: \"kubernetes.io/projected/a6363724-0c41-4b54-89e6-c672b91bbe98-kube-api-access-4c97x\") pod \"a6363724-0c41-4b54-89e6-c672b91bbe98\" (UID: \"a6363724-0c41-4b54-89e6-c672b91bbe98\") " Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.320065 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6363724-0c41-4b54-89e6-c672b91bbe98-kube-api-access-4c97x" (OuterVolumeSpecName: "kube-api-access-4c97x") pod "a6363724-0c41-4b54-89e6-c672b91bbe98" (UID: "a6363724-0c41-4b54-89e6-c672b91bbe98"). InnerVolumeSpecName "kube-api-access-4c97x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.391814 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1485-account-create-687t8" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.394016 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-95b0-account-create-txvcv" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.416142 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c97x\" (UniqueName: \"kubernetes.io/projected/a6363724-0c41-4b54-89e6-c672b91bbe98-kube-api-access-4c97x\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.517425 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5j8q\" (UniqueName: \"kubernetes.io/projected/0babfa70-91ca-45ad-afb6-aa40d2300144-kube-api-access-t5j8q\") pod \"0babfa70-91ca-45ad-afb6-aa40d2300144\" (UID: \"0babfa70-91ca-45ad-afb6-aa40d2300144\") " Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.517610 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9s2n\" (UniqueName: \"kubernetes.io/projected/1d65fbac-67f5-445f-9fd6-56ae7828fba8-kube-api-access-j9s2n\") pod \"1d65fbac-67f5-445f-9fd6-56ae7828fba8\" (UID: \"1d65fbac-67f5-445f-9fd6-56ae7828fba8\") " Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.520742 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0babfa70-91ca-45ad-afb6-aa40d2300144-kube-api-access-t5j8q" (OuterVolumeSpecName: "kube-api-access-t5j8q") pod "0babfa70-91ca-45ad-afb6-aa40d2300144" (UID: "0babfa70-91ca-45ad-afb6-aa40d2300144"). InnerVolumeSpecName "kube-api-access-t5j8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.530465 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d65fbac-67f5-445f-9fd6-56ae7828fba8-kube-api-access-j9s2n" (OuterVolumeSpecName: "kube-api-access-j9s2n") pod "1d65fbac-67f5-445f-9fd6-56ae7828fba8" (UID: "1d65fbac-67f5-445f-9fd6-56ae7828fba8"). InnerVolumeSpecName "kube-api-access-j9s2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.623113 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5j8q\" (UniqueName: \"kubernetes.io/projected/0babfa70-91ca-45ad-afb6-aa40d2300144-kube-api-access-t5j8q\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.623463 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9s2n\" (UniqueName: \"kubernetes.io/projected/1d65fbac-67f5-445f-9fd6-56ae7828fba8-kube-api-access-j9s2n\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.767452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerStarted","Data":"1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.770130 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-95b0-account-create-txvcv" event={"ID":"0babfa70-91ca-45ad-afb6-aa40d2300144","Type":"ContainerDied","Data":"160a09cb18861d3fc5c7e43e8458d2ffae624dfe4c9a58b4d7e0b579193d3148"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.770150 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-95b0-account-create-txvcv" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.770160 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="160a09cb18861d3fc5c7e43e8458d2ffae624dfe4c9a58b4d7e0b579193d3148" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.775148 4810 generic.go:334] "Generic (PLEG): container finished" podID="22742881-9836-4dbb-bd8c-deba4194d7c3" containerID="76a5e6948d4daf32096a3e2e398fd4087c887c28e01b46a3346cda1ecc593c0c" exitCode=0 Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.775226 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h7z72" event={"ID":"22742881-9836-4dbb-bd8c-deba4194d7c3","Type":"ContainerDied","Data":"76a5e6948d4daf32096a3e2e398fd4087c887c28e01b46a3346cda1ecc593c0c"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.780570 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b73c-account-create-6nkpk" event={"ID":"a6363724-0c41-4b54-89e6-c672b91bbe98","Type":"ContainerDied","Data":"6b8162095ccf49350ff0d4c931f3621aa617a20c0b5493195de3ba1f78f848cb"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.780612 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b8162095ccf49350ff0d4c931f3621aa617a20c0b5493195de3ba1f78f848cb" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.780664 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b73c-account-create-6nkpk" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.797872 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ad9084b-295c-49da-a76c-aa3cde9ffa29","Type":"ContainerStarted","Data":"829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.807429 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1485-account-create-687t8" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.807448 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1485-account-create-687t8" event={"ID":"1d65fbac-67f5-445f-9fd6-56ae7828fba8","Type":"ContainerDied","Data":"a3764bcaaa54c8ab7f76147a1eff870d8993a3530ab9f781a1da4d9e3fb7a899"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.807484 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3764bcaaa54c8ab7f76147a1eff870d8993a3530ab9f781a1da4d9e3fb7a899" Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.812425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2q9f" event={"ID":"9d3b9863-3ccd-4bda-93da-ee17ac885040","Type":"ContainerStarted","Data":"8886bbf301fede3b944ade8f43dbcf03b0120838fdcffa8ac339943a18759597"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.824217 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5632ea60-f549-438f-a244-50f811fefc1a","Type":"ContainerStarted","Data":"e2fd952082a2646936a99d8f5498c1dc178476cd9cb299a67544c11331778da9"} Oct 08 06:48:57 crc kubenswrapper[4810]: I1008 06:48:57.842168 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-m2q9f" podStartSLOduration=3.842143765 podStartE2EDuration="3.842143765s" podCreationTimestamp="2025-10-08 06:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:57.835795728 +0000 UTC m=+1040.470235468" watchObservedRunningTime="2025-10-08 06:48:57.842143765 +0000 UTC m=+1040.476583505" Oct 08 06:48:58 crc kubenswrapper[4810]: I1008 06:48:58.085052 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78b447d3-3cd5-440a-8879-02c91e23e043" path="/var/lib/kubelet/pods/78b447d3-3cd5-440a-8879-02c91e23e043/volumes" Oct 08 06:48:58 crc kubenswrapper[4810]: I1008 06:48:58.835558 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ad9084b-295c-49da-a76c-aa3cde9ffa29","Type":"ContainerStarted","Data":"986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620"} Oct 08 06:48:58 crc kubenswrapper[4810]: I1008 06:48:58.838462 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5632ea60-f549-438f-a244-50f811fefc1a","Type":"ContainerStarted","Data":"17af0d67e7c101fc6c86a1b89cffcf25c43ff30824d53b0441e8c67f457254be"} Oct 08 06:48:58 crc kubenswrapper[4810]: I1008 06:48:58.868378 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.868357908 podStartE2EDuration="4.868357908s" podCreationTimestamp="2025-10-08 06:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:58.865097228 +0000 UTC m=+1041.499536988" watchObservedRunningTime="2025-10-08 06:48:58.868357908 +0000 UTC m=+1041.502797648" Oct 08 06:48:58 crc kubenswrapper[4810]: I1008 06:48:58.897379 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.897342292 podStartE2EDuration="4.897342292s" podCreationTimestamp="2025-10-08 06:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:48:58.891567242 +0000 UTC m=+1041.526006982" watchObservedRunningTime="2025-10-08 06:48:58.897342292 +0000 UTC m=+1041.531782042" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.224126 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337277 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-7hkxc"] Oct 08 06:48:59 crc kubenswrapper[4810]: E1008 06:48:59.337614 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d65fbac-67f5-445f-9fd6-56ae7828fba8" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337628 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d65fbac-67f5-445f-9fd6-56ae7828fba8" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: E1008 06:48:59.337639 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6363724-0c41-4b54-89e6-c672b91bbe98" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337645 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6363724-0c41-4b54-89e6-c672b91bbe98" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: E1008 06:48:59.337663 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22742881-9836-4dbb-bd8c-deba4194d7c3" containerName="placement-db-sync" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337668 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="22742881-9836-4dbb-bd8c-deba4194d7c3" containerName="placement-db-sync" Oct 08 06:48:59 crc kubenswrapper[4810]: E1008 06:48:59.337683 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0babfa70-91ca-45ad-afb6-aa40d2300144" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337689 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0babfa70-91ca-45ad-afb6-aa40d2300144" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: E1008 06:48:59.337711 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b447d3-3cd5-440a-8879-02c91e23e043" containerName="init" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337716 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b447d3-3cd5-440a-8879-02c91e23e043" containerName="init" Oct 08 06:48:59 crc kubenswrapper[4810]: E1008 06:48:59.337733 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b447d3-3cd5-440a-8879-02c91e23e043" containerName="dnsmasq-dns" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337739 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b447d3-3cd5-440a-8879-02c91e23e043" containerName="dnsmasq-dns" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337893 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d65fbac-67f5-445f-9fd6-56ae7828fba8" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337904 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6363724-0c41-4b54-89e6-c672b91bbe98" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337914 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="78b447d3-3cd5-440a-8879-02c91e23e043" containerName="dnsmasq-dns" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337934 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="22742881-9836-4dbb-bd8c-deba4194d7c3" containerName="placement-db-sync" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.337944 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0babfa70-91ca-45ad-afb6-aa40d2300144" containerName="mariadb-account-create" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.338532 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.345492 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-s4ncc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.348439 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.353611 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7hkxc"] Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.373529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-scripts\") pod \"22742881-9836-4dbb-bd8c-deba4194d7c3\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.373582 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-config-data\") pod \"22742881-9836-4dbb-bd8c-deba4194d7c3\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.373635 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22742881-9836-4dbb-bd8c-deba4194d7c3-logs\") pod \"22742881-9836-4dbb-bd8c-deba4194d7c3\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.373791 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-combined-ca-bundle\") pod \"22742881-9836-4dbb-bd8c-deba4194d7c3\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.373825 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnkvs\" (UniqueName: \"kubernetes.io/projected/22742881-9836-4dbb-bd8c-deba4194d7c3-kube-api-access-qnkvs\") pod \"22742881-9836-4dbb-bd8c-deba4194d7c3\" (UID: \"22742881-9836-4dbb-bd8c-deba4194d7c3\") " Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.377239 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22742881-9836-4dbb-bd8c-deba4194d7c3-logs" (OuterVolumeSpecName: "logs") pod "22742881-9836-4dbb-bd8c-deba4194d7c3" (UID: "22742881-9836-4dbb-bd8c-deba4194d7c3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.385259 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-scripts" (OuterVolumeSpecName: "scripts") pod "22742881-9836-4dbb-bd8c-deba4194d7c3" (UID: "22742881-9836-4dbb-bd8c-deba4194d7c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.385447 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22742881-9836-4dbb-bd8c-deba4194d7c3-kube-api-access-qnkvs" (OuterVolumeSpecName: "kube-api-access-qnkvs") pod "22742881-9836-4dbb-bd8c-deba4194d7c3" (UID: "22742881-9836-4dbb-bd8c-deba4194d7c3"). InnerVolumeSpecName "kube-api-access-qnkvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.408142 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22742881-9836-4dbb-bd8c-deba4194d7c3" (UID: "22742881-9836-4dbb-bd8c-deba4194d7c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.428098 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-config-data" (OuterVolumeSpecName: "config-data") pod "22742881-9836-4dbb-bd8c-deba4194d7c3" (UID: "22742881-9836-4dbb-bd8c-deba4194d7c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476483 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx8lr\" (UniqueName: \"kubernetes.io/projected/82b5e634-87c4-4b75-9d3d-b78816226084-kube-api-access-wx8lr\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476536 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-combined-ca-bundle\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476567 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-db-sync-config-data\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476728 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476743 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnkvs\" (UniqueName: \"kubernetes.io/projected/22742881-9836-4dbb-bd8c-deba4194d7c3-kube-api-access-qnkvs\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476754 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476764 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22742881-9836-4dbb-bd8c-deba4194d7c3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.476773 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22742881-9836-4dbb-bd8c-deba4194d7c3-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.494021 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-wqx77"] Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.495316 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.497612 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9xsgq" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.497957 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.498227 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.512593 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wqx77"] Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578228 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-combined-ca-bundle\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578298 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbpv4\" (UniqueName: \"kubernetes.io/projected/cb95c243-0c64-4937-a86e-34f348023513-kube-api-access-mbpv4\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578483 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx8lr\" (UniqueName: \"kubernetes.io/projected/82b5e634-87c4-4b75-9d3d-b78816226084-kube-api-access-wx8lr\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578525 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-combined-ca-bundle\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578559 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-db-sync-config-data\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-config-data\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578727 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-db-sync-config-data\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578768 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb95c243-0c64-4937-a86e-34f348023513-etc-machine-id\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.578824 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-scripts\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.582349 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-combined-ca-bundle\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.585857 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-db-sync-config-data\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.598814 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx8lr\" (UniqueName: \"kubernetes.io/projected/82b5e634-87c4-4b75-9d3d-b78816226084-kube-api-access-wx8lr\") pod \"barbican-db-sync-7hkxc\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.663576 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.680058 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-config-data\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.680116 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-db-sync-config-data\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.680145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb95c243-0c64-4937-a86e-34f348023513-etc-machine-id\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.680173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-scripts\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.680200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-combined-ca-bundle\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.680231 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbpv4\" (UniqueName: \"kubernetes.io/projected/cb95c243-0c64-4937-a86e-34f348023513-kube-api-access-mbpv4\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.685380 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb95c243-0c64-4937-a86e-34f348023513-etc-machine-id\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.692622 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-scripts\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.693334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-config-data\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.696363 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-db-sync-config-data\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.699667 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-combined-ca-bundle\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.730683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbpv4\" (UniqueName: \"kubernetes.io/projected/cb95c243-0c64-4937-a86e-34f348023513-kube-api-access-mbpv4\") pod \"cinder-db-sync-wqx77\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.793181 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-bx46b"] Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.794705 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bx46b" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.803880 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.808636 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.813211 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lwvzc" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.816288 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bx46b"] Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.830504 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wqx77" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.886452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j69tg\" (UniqueName: \"kubernetes.io/projected/622e00b4-7ccf-4f15-b722-bf67388fdb83-kube-api-access-j69tg\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.886524 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-config\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.886633 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-combined-ca-bundle\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.932666 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h7z72" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.934070 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h7z72" event={"ID":"22742881-9836-4dbb-bd8c-deba4194d7c3","Type":"ContainerDied","Data":"0a126c6c7d59bed5a5cfee3aeead62f6837383badf81572632ead27ed551843c"} Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.934160 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a126c6c7d59bed5a5cfee3aeead62f6837383badf81572632ead27ed551843c" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.998060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-combined-ca-bundle\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:48:59 crc kubenswrapper[4810]: I1008 06:48:59.999887 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j69tg\" (UniqueName: \"kubernetes.io/projected/622e00b4-7ccf-4f15-b722-bf67388fdb83-kube-api-access-j69tg\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.000019 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-config\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.006646 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-config\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.014560 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-combined-ca-bundle\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.061255 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j69tg\" (UniqueName: \"kubernetes.io/projected/622e00b4-7ccf-4f15-b722-bf67388fdb83-kube-api-access-j69tg\") pod \"neutron-db-sync-bx46b\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.106054 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-684c4d4db6-dr9wq"] Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.108502 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.111921 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.112014 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qccns" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.112298 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.112543 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.112723 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.113383 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.126674 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-684c4d4db6-dr9wq"] Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.212268 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-scripts\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.212824 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-config-data\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.212864 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7486bd0-d5fd-4d4f-8596-705cc22237d5-logs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.212941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-combined-ca-bundle\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.233160 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-public-tls-certs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.233269 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxg6f\" (UniqueName: \"kubernetes.io/projected/d7486bd0-d5fd-4d4f-8596-705cc22237d5-kube-api-access-qxg6f\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.234091 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-internal-tls-certs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.339745 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-scripts\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.339831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-config-data\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.339890 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7486bd0-d5fd-4d4f-8596-705cc22237d5-logs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.339946 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-combined-ca-bundle\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.340006 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-public-tls-certs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.340036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxg6f\" (UniqueName: \"kubernetes.io/projected/d7486bd0-d5fd-4d4f-8596-705cc22237d5-kube-api-access-qxg6f\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.340084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-internal-tls-certs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.342100 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7486bd0-d5fd-4d4f-8596-705cc22237d5-logs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.348388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-scripts\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.351182 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-config-data\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.354272 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-public-tls-certs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.359846 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-combined-ca-bundle\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.367889 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-internal-tls-certs\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.376457 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxg6f\" (UniqueName: \"kubernetes.io/projected/d7486bd0-d5fd-4d4f-8596-705cc22237d5-kube-api-access-qxg6f\") pod \"placement-684c4d4db6-dr9wq\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.381736 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7hkxc"] Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.559507 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wqx77"] Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.566221 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.709392 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bx46b"] Oct 08 06:49:00 crc kubenswrapper[4810]: W1008 06:49:00.721900 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod622e00b4_7ccf_4f15_b722_bf67388fdb83.slice/crio-12b5633cdac1fce695fea85fd4909425dc98caea67aa157cfed94a5ffc99b99a WatchSource:0}: Error finding container 12b5633cdac1fce695fea85fd4909425dc98caea67aa157cfed94a5ffc99b99a: Status 404 returned error can't find the container with id 12b5633cdac1fce695fea85fd4909425dc98caea67aa157cfed94a5ffc99b99a Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.943205 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bx46b" event={"ID":"622e00b4-7ccf-4f15-b722-bf67388fdb83","Type":"ContainerStarted","Data":"7d46128d5daaaf86caffd08803a49b7c5f8c2451b094dba375d4a115f05b3abd"} Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.943664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bx46b" event={"ID":"622e00b4-7ccf-4f15-b722-bf67388fdb83","Type":"ContainerStarted","Data":"12b5633cdac1fce695fea85fd4909425dc98caea67aa157cfed94a5ffc99b99a"} Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.948570 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wqx77" event={"ID":"cb95c243-0c64-4937-a86e-34f348023513","Type":"ContainerStarted","Data":"c70e42e3f471ef7f555620f357e1f14834d3321f897ce4e89df12bf910317d62"} Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.951311 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7hkxc" event={"ID":"82b5e634-87c4-4b75-9d3d-b78816226084","Type":"ContainerStarted","Data":"2adf0879832c71b7b571578b83a5c3d6ae215a8014a700044a322462539fb4cf"} Oct 08 06:49:00 crc kubenswrapper[4810]: I1008 06:49:00.961121 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-bx46b" podStartSLOduration=1.961100322 podStartE2EDuration="1.961100322s" podCreationTimestamp="2025-10-08 06:48:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:00.959806557 +0000 UTC m=+1043.594246307" watchObservedRunningTime="2025-10-08 06:49:00.961100322 +0000 UTC m=+1043.595540052" Oct 08 06:49:01 crc kubenswrapper[4810]: I1008 06:49:01.075818 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-684c4d4db6-dr9wq"] Oct 08 06:49:01 crc kubenswrapper[4810]: I1008 06:49:01.963796 4810 generic.go:334] "Generic (PLEG): container finished" podID="9d3b9863-3ccd-4bda-93da-ee17ac885040" containerID="8886bbf301fede3b944ade8f43dbcf03b0120838fdcffa8ac339943a18759597" exitCode=0 Oct 08 06:49:01 crc kubenswrapper[4810]: I1008 06:49:01.963898 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2q9f" event={"ID":"9d3b9863-3ccd-4bda-93da-ee17ac885040","Type":"ContainerDied","Data":"8886bbf301fede3b944ade8f43dbcf03b0120838fdcffa8ac339943a18759597"} Oct 08 06:49:05 crc kubenswrapper[4810]: W1008 06:49:05.120102 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7486bd0_d5fd_4d4f_8596_705cc22237d5.slice/crio-cb36ae6379de91ab949ca3be579c93e33e60abbc00d6023c61489cd41a69f8bf WatchSource:0}: Error finding container cb36ae6379de91ab949ca3be579c93e33e60abbc00d6023c61489cd41a69f8bf: Status 404 returned error can't find the container with id cb36ae6379de91ab949ca3be579c93e33e60abbc00d6023c61489cd41a69f8bf Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.323703 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.324172 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.348899 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.433556 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.435334 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.514466 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ktf2\" (UniqueName: \"kubernetes.io/projected/9d3b9863-3ccd-4bda-93da-ee17ac885040-kube-api-access-2ktf2\") pod \"9d3b9863-3ccd-4bda-93da-ee17ac885040\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.514532 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-scripts\") pod \"9d3b9863-3ccd-4bda-93da-ee17ac885040\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.514629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-credential-keys\") pod \"9d3b9863-3ccd-4bda-93da-ee17ac885040\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.514676 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-config-data\") pod \"9d3b9863-3ccd-4bda-93da-ee17ac885040\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.514747 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-fernet-keys\") pod \"9d3b9863-3ccd-4bda-93da-ee17ac885040\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.514824 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-combined-ca-bundle\") pod \"9d3b9863-3ccd-4bda-93da-ee17ac885040\" (UID: \"9d3b9863-3ccd-4bda-93da-ee17ac885040\") " Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.522803 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-scripts" (OuterVolumeSpecName: "scripts") pod "9d3b9863-3ccd-4bda-93da-ee17ac885040" (UID: "9d3b9863-3ccd-4bda-93da-ee17ac885040"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.522887 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9d3b9863-3ccd-4bda-93da-ee17ac885040" (UID: "9d3b9863-3ccd-4bda-93da-ee17ac885040"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.523120 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3b9863-3ccd-4bda-93da-ee17ac885040-kube-api-access-2ktf2" (OuterVolumeSpecName: "kube-api-access-2ktf2") pod "9d3b9863-3ccd-4bda-93da-ee17ac885040" (UID: "9d3b9863-3ccd-4bda-93da-ee17ac885040"). InnerVolumeSpecName "kube-api-access-2ktf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.527215 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9d3b9863-3ccd-4bda-93da-ee17ac885040" (UID: "9d3b9863-3ccd-4bda-93da-ee17ac885040"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.545993 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-config-data" (OuterVolumeSpecName: "config-data") pod "9d3b9863-3ccd-4bda-93da-ee17ac885040" (UID: "9d3b9863-3ccd-4bda-93da-ee17ac885040"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.549743 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d3b9863-3ccd-4bda-93da-ee17ac885040" (UID: "9d3b9863-3ccd-4bda-93da-ee17ac885040"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.601681 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.603293 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.617242 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ktf2\" (UniqueName: \"kubernetes.io/projected/9d3b9863-3ccd-4bda-93da-ee17ac885040-kube-api-access-2ktf2\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.617294 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.617312 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.617324 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.617338 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.617351 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3b9863-3ccd-4bda-93da-ee17ac885040-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.661518 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 06:49:05 crc kubenswrapper[4810]: I1008 06:49:05.675778 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.016443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerStarted","Data":"533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec"} Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.021776 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684c4d4db6-dr9wq" event={"ID":"d7486bd0-d5fd-4d4f-8596-705cc22237d5","Type":"ContainerStarted","Data":"b526413b9a8c8e6cae317ae5e79f9dced75f63320d48b38b7a09dab4a2bdcb30"} Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.021825 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684c4d4db6-dr9wq" event={"ID":"d7486bd0-d5fd-4d4f-8596-705cc22237d5","Type":"ContainerStarted","Data":"3556449533e62b8a1b045f6331743feae434169e2723d25a19669ad4c27722e9"} Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.021838 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684c4d4db6-dr9wq" event={"ID":"d7486bd0-d5fd-4d4f-8596-705cc22237d5","Type":"ContainerStarted","Data":"cb36ae6379de91ab949ca3be579c93e33e60abbc00d6023c61489cd41a69f8bf"} Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.021882 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.033509 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2q9f" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.033622 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2q9f" event={"ID":"9d3b9863-3ccd-4bda-93da-ee17ac885040","Type":"ContainerDied","Data":"8744b5ebe7a4650dca24b3baf7106f29051ba778b3d32ce3ffed70d395e9a149"} Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.033675 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8744b5ebe7a4650dca24b3baf7106f29051ba778b3d32ce3ffed70d395e9a149" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.033725 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.034864 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.034884 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.034897 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.069530 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-684c4d4db6-dr9wq" podStartSLOduration=6.06950957 podStartE2EDuration="6.06950957s" podCreationTimestamp="2025-10-08 06:49:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:06.049120263 +0000 UTC m=+1048.683560023" watchObservedRunningTime="2025-10-08 06:49:06.06950957 +0000 UTC m=+1048.703949310" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.535337 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6b4dd4748f-7h7bc"] Oct 08 06:49:06 crc kubenswrapper[4810]: E1008 06:49:06.536957 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3b9863-3ccd-4bda-93da-ee17ac885040" containerName="keystone-bootstrap" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.536981 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3b9863-3ccd-4bda-93da-ee17ac885040" containerName="keystone-bootstrap" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.537264 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3b9863-3ccd-4bda-93da-ee17ac885040" containerName="keystone-bootstrap" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.540191 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b4dd4748f-7h7bc"] Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.540277 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.543187 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.543405 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.545346 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.545510 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tb5ft" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.545632 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.545841 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.643761 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.644245 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-config-data\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.644292 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-scripts\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.644350 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-fernet-keys\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.644397 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-public-tls-certs\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.644428 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt45m\" (UniqueName: \"kubernetes.io/projected/493810ad-f07f-4871-8063-5d54cd5786ea-kube-api-access-rt45m\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.644457 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-credential-keys\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.644493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-combined-ca-bundle\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.746458 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.746518 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-config-data\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.746559 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-scripts\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.746634 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-fernet-keys\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.747664 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-public-tls-certs\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.747698 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt45m\" (UniqueName: \"kubernetes.io/projected/493810ad-f07f-4871-8063-5d54cd5786ea-kube-api-access-rt45m\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.747725 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-credential-keys\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.747762 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-combined-ca-bundle\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.756147 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-public-tls-certs\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.759907 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-config-data\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.761582 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.761645 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-scripts\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.765070 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-credential-keys\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.765604 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-fernet-keys\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.765928 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-combined-ca-bundle\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.783652 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt45m\" (UniqueName: \"kubernetes.io/projected/493810ad-f07f-4871-8063-5d54cd5786ea-kube-api-access-rt45m\") pod \"keystone-6b4dd4748f-7h7bc\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:06 crc kubenswrapper[4810]: I1008 06:49:06.860802 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:07 crc kubenswrapper[4810]: I1008 06:49:07.052974 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:07 crc kubenswrapper[4810]: I1008 06:49:07.500144 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b4dd4748f-7h7bc"] Oct 08 06:49:07 crc kubenswrapper[4810]: W1008 06:49:07.525772 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod493810ad_f07f_4871_8063_5d54cd5786ea.slice/crio-c18483cfeeab72004f52e63b3aa0a878e963ab4fa03645b5570b9c2dd3b3aa1f WatchSource:0}: Error finding container c18483cfeeab72004f52e63b3aa0a878e963ab4fa03645b5570b9c2dd3b3aa1f: Status 404 returned error can't find the container with id c18483cfeeab72004f52e63b3aa0a878e963ab4fa03645b5570b9c2dd3b3aa1f Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.065661 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b4dd4748f-7h7bc" event={"ID":"493810ad-f07f-4871-8063-5d54cd5786ea","Type":"ContainerStarted","Data":"5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e"} Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.065933 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.066048 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.066104 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.066183 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.066164 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.066211 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b4dd4748f-7h7bc" event={"ID":"493810ad-f07f-4871-8063-5d54cd5786ea","Type":"ContainerStarted","Data":"c18483cfeeab72004f52e63b3aa0a878e963ab4fa03645b5570b9c2dd3b3aa1f"} Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.095872 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6b4dd4748f-7h7bc" podStartSLOduration=2.095843081 podStartE2EDuration="2.095843081s" podCreationTimestamp="2025-10-08 06:49:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:08.08390301 +0000 UTC m=+1050.718342750" watchObservedRunningTime="2025-10-08 06:49:08.095843081 +0000 UTC m=+1050.730282831" Oct 08 06:49:08 crc kubenswrapper[4810]: I1008 06:49:08.752028 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:09 crc kubenswrapper[4810]: I1008 06:49:09.082245 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:49:09 crc kubenswrapper[4810]: I1008 06:49:09.175411 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 06:49:09 crc kubenswrapper[4810]: I1008 06:49:09.175553 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:49:09 crc kubenswrapper[4810]: I1008 06:49:09.177620 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 06:49:09 crc kubenswrapper[4810]: I1008 06:49:09.356238 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 06:49:23 crc kubenswrapper[4810]: I1008 06:49:23.277530 4810 generic.go:334] "Generic (PLEG): container finished" podID="622e00b4-7ccf-4f15-b722-bf67388fdb83" containerID="7d46128d5daaaf86caffd08803a49b7c5f8c2451b094dba375d4a115f05b3abd" exitCode=0 Oct 08 06:49:23 crc kubenswrapper[4810]: I1008 06:49:23.277620 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bx46b" event={"ID":"622e00b4-7ccf-4f15-b722-bf67388fdb83","Type":"ContainerDied","Data":"7d46128d5daaaf86caffd08803a49b7c5f8c2451b094dba375d4a115f05b3abd"} Oct 08 06:49:25 crc kubenswrapper[4810]: E1008 06:49:25.393373 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 08 06:49:25 crc kubenswrapper[4810]: E1008 06:49:25.394589 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mbpv4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-wqx77_openstack(cb95c243-0c64-4937-a86e-34f348023513): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 06:49:25 crc kubenswrapper[4810]: E1008 06:49:25.395911 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-wqx77" podUID="cb95c243-0c64-4937-a86e-34f348023513" Oct 08 06:49:25 crc kubenswrapper[4810]: E1008 06:49:25.785312 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 08 06:49:25 crc kubenswrapper[4810]: E1008 06:49:25.785512 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nc5xm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(6b854adc-e2af-4b4e-a842-985cb1ffbb2a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 06:49:25 crc kubenswrapper[4810]: E1008 06:49:25.787270 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" Oct 08 06:49:25 crc kubenswrapper[4810]: I1008 06:49:25.926547 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.107936 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-config\") pod \"622e00b4-7ccf-4f15-b722-bf67388fdb83\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.108427 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j69tg\" (UniqueName: \"kubernetes.io/projected/622e00b4-7ccf-4f15-b722-bf67388fdb83-kube-api-access-j69tg\") pod \"622e00b4-7ccf-4f15-b722-bf67388fdb83\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.108471 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-combined-ca-bundle\") pod \"622e00b4-7ccf-4f15-b722-bf67388fdb83\" (UID: \"622e00b4-7ccf-4f15-b722-bf67388fdb83\") " Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.114874 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/622e00b4-7ccf-4f15-b722-bf67388fdb83-kube-api-access-j69tg" (OuterVolumeSpecName: "kube-api-access-j69tg") pod "622e00b4-7ccf-4f15-b722-bf67388fdb83" (UID: "622e00b4-7ccf-4f15-b722-bf67388fdb83"). InnerVolumeSpecName "kube-api-access-j69tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.136813 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "622e00b4-7ccf-4f15-b722-bf67388fdb83" (UID: "622e00b4-7ccf-4f15-b722-bf67388fdb83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.144465 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-config" (OuterVolumeSpecName: "config") pod "622e00b4-7ccf-4f15-b722-bf67388fdb83" (UID: "622e00b4-7ccf-4f15-b722-bf67388fdb83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.213714 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j69tg\" (UniqueName: \"kubernetes.io/projected/622e00b4-7ccf-4f15-b722-bf67388fdb83-kube-api-access-j69tg\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.213767 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.213788 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/622e00b4-7ccf-4f15-b722-bf67388fdb83-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.321504 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7hkxc" event={"ID":"82b5e634-87c4-4b75-9d3d-b78816226084","Type":"ContainerStarted","Data":"66a0ef3b680887a96f474c84ba92ef0028c19cf0633bc4c7f831bb1f67bcc308"} Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.329845 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bx46b" event={"ID":"622e00b4-7ccf-4f15-b722-bf67388fdb83","Type":"ContainerDied","Data":"12b5633cdac1fce695fea85fd4909425dc98caea67aa157cfed94a5ffc99b99a"} Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.329904 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12b5633cdac1fce695fea85fd4909425dc98caea67aa157cfed94a5ffc99b99a" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.339162 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bx46b" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.340697 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-central-agent" containerID="cri-o://8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66" gracePeriod=30 Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.340922 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="sg-core" containerID="cri-o://533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec" gracePeriod=30 Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.341038 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-notification-agent" containerID="cri-o://1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039" gracePeriod=30 Oct 08 06:49:26 crc kubenswrapper[4810]: E1008 06:49:26.341515 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-wqx77" podUID="cb95c243-0c64-4937-a86e-34f348023513" Oct 08 06:49:26 crc kubenswrapper[4810]: I1008 06:49:26.357130 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-7hkxc" podStartSLOduration=2.005624578 podStartE2EDuration="27.357093852s" podCreationTimestamp="2025-10-08 06:48:59 +0000 UTC" firstStartedPulling="2025-10-08 06:49:00.421537182 +0000 UTC m=+1043.055976922" lastFinishedPulling="2025-10-08 06:49:25.773006456 +0000 UTC m=+1068.407446196" observedRunningTime="2025-10-08 06:49:26.350657783 +0000 UTC m=+1068.985097553" watchObservedRunningTime="2025-10-08 06:49:26.357093852 +0000 UTC m=+1068.991533632" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.226487 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b55c5465-xfkrq"] Oct 08 06:49:27 crc kubenswrapper[4810]: E1008 06:49:27.227490 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="622e00b4-7ccf-4f15-b722-bf67388fdb83" containerName="neutron-db-sync" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.227512 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="622e00b4-7ccf-4f15-b722-bf67388fdb83" containerName="neutron-db-sync" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.227987 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="622e00b4-7ccf-4f15-b722-bf67388fdb83" containerName="neutron-db-sync" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.235682 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.263402 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b55c5465-xfkrq"] Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.346531 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cs68\" (UniqueName: \"kubernetes.io/projected/20e53f85-d85c-466f-84b5-373d6d6af76b-kube-api-access-2cs68\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.346635 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-swift-storage-0\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.346717 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-svc\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.346746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-config\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.346772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-sb\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.346801 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-nb\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.371669 4810 generic.go:334] "Generic (PLEG): container finished" podID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerID="533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec" exitCode=2 Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.371715 4810 generic.go:334] "Generic (PLEG): container finished" podID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerID="8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66" exitCode=0 Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.372067 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerDied","Data":"533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec"} Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.372101 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerDied","Data":"8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66"} Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.383342 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-67b6b58864-6qdbr"] Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.385351 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.389209 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.389356 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.395003 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lwvzc" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.395436 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.402135 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67b6b58864-6qdbr"] Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.448077 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-svc\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.448145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-config\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.448195 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-sb\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.448241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-nb\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.448301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cs68\" (UniqueName: \"kubernetes.io/projected/20e53f85-d85c-466f-84b5-373d6d6af76b-kube-api-access-2cs68\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.448370 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-swift-storage-0\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.449751 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-swift-storage-0\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.450189 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-config\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.450566 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-sb\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.450951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-nb\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.452406 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-svc\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.470001 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cs68\" (UniqueName: \"kubernetes.io/projected/20e53f85-d85c-466f-84b5-373d6d6af76b-kube-api-access-2cs68\") pod \"dnsmasq-dns-67b55c5465-xfkrq\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.550225 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-ovndb-tls-certs\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.550413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-config\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.550452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-combined-ca-bundle\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.550470 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-httpd-config\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.550526 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8psk\" (UniqueName: \"kubernetes.io/projected/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-kube-api-access-q8psk\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.570557 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.651922 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-config\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.652341 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-httpd-config\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.652365 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-combined-ca-bundle\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.652422 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8psk\" (UniqueName: \"kubernetes.io/projected/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-kube-api-access-q8psk\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.652465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-ovndb-tls-certs\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.660486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-ovndb-tls-certs\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.660716 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-config\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.665747 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-httpd-config\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.667624 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-combined-ca-bundle\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.683945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8psk\" (UniqueName: \"kubernetes.io/projected/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-kube-api-access-q8psk\") pod \"neutron-67b6b58864-6qdbr\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.711566 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:27 crc kubenswrapper[4810]: I1008 06:49:27.943444 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b55c5465-xfkrq"] Oct 08 06:49:28 crc kubenswrapper[4810]: I1008 06:49:28.144502 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67b6b58864-6qdbr"] Oct 08 06:49:28 crc kubenswrapper[4810]: W1008 06:49:28.151995 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1edfb10_afed_4da0_aa22_9a1e0ab2c22a.slice/crio-f4825f3141c2fd29848bc9240bb7bb135baf7c08e0e2c4c042838ffcf3a1073c WatchSource:0}: Error finding container f4825f3141c2fd29848bc9240bb7bb135baf7c08e0e2c4c042838ffcf3a1073c: Status 404 returned error can't find the container with id f4825f3141c2fd29848bc9240bb7bb135baf7c08e0e2c4c042838ffcf3a1073c Oct 08 06:49:28 crc kubenswrapper[4810]: I1008 06:49:28.383564 4810 generic.go:334] "Generic (PLEG): container finished" podID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerID="6501f4b4452bb73d358f200ef3d7c9bc4f681eb90be955d42bf54c0341695bbe" exitCode=0 Oct 08 06:49:28 crc kubenswrapper[4810]: I1008 06:49:28.383669 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" event={"ID":"20e53f85-d85c-466f-84b5-373d6d6af76b","Type":"ContainerDied","Data":"6501f4b4452bb73d358f200ef3d7c9bc4f681eb90be955d42bf54c0341695bbe"} Oct 08 06:49:28 crc kubenswrapper[4810]: I1008 06:49:28.384182 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" event={"ID":"20e53f85-d85c-466f-84b5-373d6d6af76b","Type":"ContainerStarted","Data":"07d5ca0d0e67c7e578245d6c151d4f2346161a013685b2e9617411cf6d02999a"} Oct 08 06:49:28 crc kubenswrapper[4810]: I1008 06:49:28.386511 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67b6b58864-6qdbr" event={"ID":"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a","Type":"ContainerStarted","Data":"f4825f3141c2fd29848bc9240bb7bb135baf7c08e0e2c4c042838ffcf3a1073c"} Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.398649 4810 generic.go:334] "Generic (PLEG): container finished" podID="82b5e634-87c4-4b75-9d3d-b78816226084" containerID="66a0ef3b680887a96f474c84ba92ef0028c19cf0633bc4c7f831bb1f67bcc308" exitCode=0 Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.398756 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7hkxc" event={"ID":"82b5e634-87c4-4b75-9d3d-b78816226084","Type":"ContainerDied","Data":"66a0ef3b680887a96f474c84ba92ef0028c19cf0633bc4c7f831bb1f67bcc308"} Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.402583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" event={"ID":"20e53f85-d85c-466f-84b5-373d6d6af76b","Type":"ContainerStarted","Data":"bf6d696953c6d7d34eb261bbfc2d0cec27de8b37a20463601d0cd4f3c9ec1703"} Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.402701 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.405197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67b6b58864-6qdbr" event={"ID":"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a","Type":"ContainerStarted","Data":"f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4"} Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.405233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67b6b58864-6qdbr" event={"ID":"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a","Type":"ContainerStarted","Data":"f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf"} Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.405354 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.450300 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" podStartSLOduration=2.450276865 podStartE2EDuration="2.450276865s" podCreationTimestamp="2025-10-08 06:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:29.440506343 +0000 UTC m=+1072.074946103" watchObservedRunningTime="2025-10-08 06:49:29.450276865 +0000 UTC m=+1072.084716615" Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.475655 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-67b6b58864-6qdbr" podStartSLOduration=2.475629128 podStartE2EDuration="2.475629128s" podCreationTimestamp="2025-10-08 06:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:29.464447038 +0000 UTC m=+1072.098886778" watchObservedRunningTime="2025-10-08 06:49:29.475629128 +0000 UTC m=+1072.110068868" Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.977356 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6f65ff9-7hhz2"] Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.979340 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.985922 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.986158 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 08 06:49:29 crc kubenswrapper[4810]: I1008 06:49:29.991473 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f65ff9-7hhz2"] Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.105380 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-httpd-config\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.105438 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk47z\" (UniqueName: \"kubernetes.io/projected/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-kube-api-access-hk47z\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.106044 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-public-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.106206 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-combined-ca-bundle\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.108402 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-internal-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.108477 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-ovndb-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.108516 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-config\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.209854 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-internal-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.210190 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-ovndb-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.210266 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-config\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.210371 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-httpd-config\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.210467 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk47z\" (UniqueName: \"kubernetes.io/projected/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-kube-api-access-hk47z\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.211866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-public-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.212562 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-combined-ca-bundle\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.224018 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-httpd-config\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.224738 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-public-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.225034 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-ovndb-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.228461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-config\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.229095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-combined-ca-bundle\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.229152 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-internal-tls-certs\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.256849 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk47z\" (UniqueName: \"kubernetes.io/projected/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-kube-api-access-hk47z\") pod \"neutron-6f65ff9-7hhz2\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.303640 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.819322 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.933004 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-db-sync-config-data\") pod \"82b5e634-87c4-4b75-9d3d-b78816226084\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.933065 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx8lr\" (UniqueName: \"kubernetes.io/projected/82b5e634-87c4-4b75-9d3d-b78816226084-kube-api-access-wx8lr\") pod \"82b5e634-87c4-4b75-9d3d-b78816226084\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.933303 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-combined-ca-bundle\") pod \"82b5e634-87c4-4b75-9d3d-b78816226084\" (UID: \"82b5e634-87c4-4b75-9d3d-b78816226084\") " Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.942716 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "82b5e634-87c4-4b75-9d3d-b78816226084" (UID: "82b5e634-87c4-4b75-9d3d-b78816226084"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.942748 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b5e634-87c4-4b75-9d3d-b78816226084-kube-api-access-wx8lr" (OuterVolumeSpecName: "kube-api-access-wx8lr") pod "82b5e634-87c4-4b75-9d3d-b78816226084" (UID: "82b5e634-87c4-4b75-9d3d-b78816226084"). InnerVolumeSpecName "kube-api-access-wx8lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:30 crc kubenswrapper[4810]: I1008 06:49:30.978129 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82b5e634-87c4-4b75-9d3d-b78816226084" (UID: "82b5e634-87c4-4b75-9d3d-b78816226084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.044752 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.045315 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx8lr\" (UniqueName: \"kubernetes.io/projected/82b5e634-87c4-4b75-9d3d-b78816226084-kube-api-access-wx8lr\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.045332 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b5e634-87c4-4b75-9d3d-b78816226084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.052748 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f65ff9-7hhz2"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.130513 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.250098 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-scripts\") pod \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.250468 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-combined-ca-bundle\") pod \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.250654 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-log-httpd\") pod \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.250812 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-config-data\") pod \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.251095 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-run-httpd\") pod \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.251310 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc5xm\" (UniqueName: \"kubernetes.io/projected/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-kube-api-access-nc5xm\") pod \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.251480 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-sg-core-conf-yaml\") pod \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\" (UID: \"6b854adc-e2af-4b4e-a842-985cb1ffbb2a\") " Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.251550 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6b854adc-e2af-4b4e-a842-985cb1ffbb2a" (UID: "6b854adc-e2af-4b4e-a842-985cb1ffbb2a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.252295 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.252387 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6b854adc-e2af-4b4e-a842-985cb1ffbb2a" (UID: "6b854adc-e2af-4b4e-a842-985cb1ffbb2a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.256904 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-scripts" (OuterVolumeSpecName: "scripts") pod "6b854adc-e2af-4b4e-a842-985cb1ffbb2a" (UID: "6b854adc-e2af-4b4e-a842-985cb1ffbb2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.259375 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-kube-api-access-nc5xm" (OuterVolumeSpecName: "kube-api-access-nc5xm") pod "6b854adc-e2af-4b4e-a842-985cb1ffbb2a" (UID: "6b854adc-e2af-4b4e-a842-985cb1ffbb2a"). InnerVolumeSpecName "kube-api-access-nc5xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.298350 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6b854adc-e2af-4b4e-a842-985cb1ffbb2a" (UID: "6b854adc-e2af-4b4e-a842-985cb1ffbb2a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.316844 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b854adc-e2af-4b4e-a842-985cb1ffbb2a" (UID: "6b854adc-e2af-4b4e-a842-985cb1ffbb2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.336206 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-config-data" (OuterVolumeSpecName: "config-data") pod "6b854adc-e2af-4b4e-a842-985cb1ffbb2a" (UID: "6b854adc-e2af-4b4e-a842-985cb1ffbb2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.355161 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc5xm\" (UniqueName: \"kubernetes.io/projected/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-kube-api-access-nc5xm\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.355267 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.355288 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.355330 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.355343 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.355417 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b854adc-e2af-4b4e-a842-985cb1ffbb2a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.470105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f65ff9-7hhz2" event={"ID":"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0","Type":"ContainerStarted","Data":"77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd"} Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.470483 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f65ff9-7hhz2" event={"ID":"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0","Type":"ContainerStarted","Data":"d131e54d6c7d758c981b65b803d5b97e5dd547f70afde36f4dcf40a8ff1cae31"} Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.473082 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7hkxc" event={"ID":"82b5e634-87c4-4b75-9d3d-b78816226084","Type":"ContainerDied","Data":"2adf0879832c71b7b571578b83a5c3d6ae215a8014a700044a322462539fb4cf"} Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.473106 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2adf0879832c71b7b571578b83a5c3d6ae215a8014a700044a322462539fb4cf" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.473111 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7hkxc" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.475710 4810 generic.go:334] "Generic (PLEG): container finished" podID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerID="1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039" exitCode=0 Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.475734 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerDied","Data":"1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039"} Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.475756 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b854adc-e2af-4b4e-a842-985cb1ffbb2a","Type":"ContainerDied","Data":"b5613fc4b756fcf92f58fee50d7c007c1e163f111c308da8ffefba505019b8f3"} Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.475775 4810 scope.go:117] "RemoveContainer" containerID="533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.475812 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.520498 4810 scope.go:117] "RemoveContainer" containerID="1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.602868 4810 scope.go:117] "RemoveContainer" containerID="8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.604077 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.627658 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653072 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:31 crc kubenswrapper[4810]: E1008 06:49:31.653576 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="sg-core" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653593 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="sg-core" Oct 08 06:49:31 crc kubenswrapper[4810]: E1008 06:49:31.653603 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-notification-agent" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653609 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-notification-agent" Oct 08 06:49:31 crc kubenswrapper[4810]: E1008 06:49:31.653629 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b5e634-87c4-4b75-9d3d-b78816226084" containerName="barbican-db-sync" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653637 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b5e634-87c4-4b75-9d3d-b78816226084" containerName="barbican-db-sync" Oct 08 06:49:31 crc kubenswrapper[4810]: E1008 06:49:31.653655 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-central-agent" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653664 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-central-agent" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653854 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-notification-agent" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653870 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="ceilometer-central-agent" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653884 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" containerName="sg-core" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.653892 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b5e634-87c4-4b75-9d3d-b78816226084" containerName="barbican-db-sync" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.656747 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.660894 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.661307 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.668348 4810 scope.go:117] "RemoveContainer" containerID="533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.672326 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:31 crc kubenswrapper[4810]: E1008 06:49:31.674791 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec\": container with ID starting with 533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec not found: ID does not exist" containerID="533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.674850 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec"} err="failed to get container status \"533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec\": rpc error: code = NotFound desc = could not find container \"533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec\": container with ID starting with 533ce2c93b76b35a8ed3008f5174c9706a37ae0155e23f2ad68a91fbae7463ec not found: ID does not exist" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.674889 4810 scope.go:117] "RemoveContainer" containerID="1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039" Oct 08 06:49:31 crc kubenswrapper[4810]: E1008 06:49:31.686033 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039\": container with ID starting with 1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039 not found: ID does not exist" containerID="1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.686083 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039"} err="failed to get container status \"1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039\": rpc error: code = NotFound desc = could not find container \"1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039\": container with ID starting with 1ae0f68eb66011c159e10f3160fd820c6c794def709c0c15b4ca73800d61e039 not found: ID does not exist" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.686113 4810 scope.go:117] "RemoveContainer" containerID="8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.686722 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5b4bcc558b-zx56q"] Oct 08 06:49:31 crc kubenswrapper[4810]: E1008 06:49:31.693790 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66\": container with ID starting with 8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66 not found: ID does not exist" containerID="8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.693838 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66"} err="failed to get container status \"8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66\": rpc error: code = NotFound desc = could not find container \"8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66\": container with ID starting with 8da81d71477a5ed38e5ae1290b19e009996704c91ec24b4443caf903d83c5c66 not found: ID does not exist" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.699095 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.713649 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5b4bcc558b-zx56q"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.714534 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-s4ncc" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.714687 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.715062 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.744265 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-59f9c4fc59-kvs2g"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.745888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.756950 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.767022 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.767486 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql526\" (UniqueName: \"kubernetes.io/projected/da3b3c98-166f-433f-8e3b-38125745f14f-kube-api-access-ql526\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.767582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-run-httpd\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.767743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-config-data\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.767881 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-log-httpd\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.768013 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-scripts\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.768114 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.779343 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-59f9c4fc59-kvs2g"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.876136 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b55c5465-xfkrq"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.877109 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" podUID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerName="dnsmasq-dns" containerID="cri-o://bf6d696953c6d7d34eb261bbfc2d0cec27de8b37a20463601d0cd4f3c9ec1703" gracePeriod=10 Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.906911 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9djf\" (UniqueName: \"kubernetes.io/projected/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-kube-api-access-k9djf\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.907005 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-run-httpd\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.907030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data-custom\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.907086 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-combined-ca-bundle\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.907160 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbdqp\" (UniqueName: \"kubernetes.io/projected/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-kube-api-access-tbdqp\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.907257 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.907284 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-config-data\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.907342 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data-custom\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.913259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-run-httpd\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.944274 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-log-httpd\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.944499 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.944727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-scripts\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.944877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.945036 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-logs\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.945149 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-log-httpd\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.945230 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.945586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql526\" (UniqueName: \"kubernetes.io/projected/da3b3c98-166f-433f-8e3b-38125745f14f-kube-api-access-ql526\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.946627 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-combined-ca-bundle\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.946721 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-logs\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.970187 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c78787df7-xcmqq"] Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.971401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-config-data\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:31 crc kubenswrapper[4810]: I1008 06:49:31.987952 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql526\" (UniqueName: \"kubernetes.io/projected/da3b3c98-166f-433f-8e3b-38125745f14f-kube-api-access-ql526\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.007138 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.021069 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.022322 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c78787df7-xcmqq"] Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.023482 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-scripts\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.025179 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " pod="openstack/ceilometer-0" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.041109 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f98c8c4f6-259sf"] Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.043634 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.049015 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f98c8c4f6-259sf"] Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.049222 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-combined-ca-bundle\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050253 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-logs\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9djf\" (UniqueName: \"kubernetes.io/projected/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-kube-api-access-k9djf\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050333 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data-custom\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050376 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-combined-ca-bundle\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050420 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbdqp\" (UniqueName: \"kubernetes.io/projected/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-kube-api-access-tbdqp\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050540 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data-custom\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050612 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.050701 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-logs\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.051246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-logs\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.056286 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-combined-ca-bundle\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.057585 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data-custom\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.060490 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-logs\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.063943 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.064816 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data-custom\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.066281 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.073059 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-combined-ca-bundle\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.078861 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbdqp\" (UniqueName: \"kubernetes.io/projected/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-kube-api-access-tbdqp\") pod \"barbican-worker-59f9c4fc59-kvs2g\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.079983 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.085344 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9djf\" (UniqueName: \"kubernetes.io/projected/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-kube-api-access-k9djf\") pod \"barbican-keystone-listener-5b4bcc558b-zx56q\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.103850 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b854adc-e2af-4b4e-a842-985cb1ffbb2a" path="/var/lib/kubelet/pods/6b854adc-e2af-4b4e-a842-985cb1ffbb2a/volumes" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.108541 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-combined-ca-bundle\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-svc\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152278 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv4ck\" (UniqueName: \"kubernetes.io/projected/791f9b40-309e-4266-9962-919281a31ffa-kube-api-access-hv4ck\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152301 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data-custom\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152320 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-swift-storage-0\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152349 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp465\" (UniqueName: \"kubernetes.io/projected/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-kube-api-access-vp465\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152385 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-logs\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152409 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-config\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152475 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-sb\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.152493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-nb\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.194304 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.259499 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp465\" (UniqueName: \"kubernetes.io/projected/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-kube-api-access-vp465\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.259598 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-logs\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260232 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-config\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260267 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-sb\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260286 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-nb\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260342 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-combined-ca-bundle\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260371 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-svc\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv4ck\" (UniqueName: \"kubernetes.io/projected/791f9b40-309e-4266-9962-919281a31ffa-kube-api-access-hv4ck\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260448 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data-custom\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.260479 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-swift-storage-0\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.261314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-logs\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.261748 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-swift-storage-0\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.264138 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-nb\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.265616 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-svc\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.266250 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-config\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.267084 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-sb\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.274580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.281776 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data-custom\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.289951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv4ck\" (UniqueName: \"kubernetes.io/projected/791f9b40-309e-4266-9962-919281a31ffa-kube-api-access-hv4ck\") pod \"dnsmasq-dns-5c78787df7-xcmqq\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.291018 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-combined-ca-bundle\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.293765 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp465\" (UniqueName: \"kubernetes.io/projected/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-kube-api-access-vp465\") pod \"barbican-api-6f98c8c4f6-259sf\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.322101 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.342635 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.376633 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.392620 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.519481 4810 generic.go:334] "Generic (PLEG): container finished" podID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerID="bf6d696953c6d7d34eb261bbfc2d0cec27de8b37a20463601d0cd4f3c9ec1703" exitCode=0 Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.519575 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" event={"ID":"20e53f85-d85c-466f-84b5-373d6d6af76b","Type":"ContainerDied","Data":"bf6d696953c6d7d34eb261bbfc2d0cec27de8b37a20463601d0cd4f3c9ec1703"} Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.529208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f65ff9-7hhz2" event={"ID":"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0","Type":"ContainerStarted","Data":"a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df"} Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.529265 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.533165 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.576506 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6f65ff9-7hhz2" podStartSLOduration=3.5764762230000002 podStartE2EDuration="3.576476223s" podCreationTimestamp="2025-10-08 06:49:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:32.551915992 +0000 UTC m=+1075.186355732" watchObservedRunningTime="2025-10-08 06:49:32.576476223 +0000 UTC m=+1075.210915963" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.672405 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cs68\" (UniqueName: \"kubernetes.io/projected/20e53f85-d85c-466f-84b5-373d6d6af76b-kube-api-access-2cs68\") pod \"20e53f85-d85c-466f-84b5-373d6d6af76b\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.672541 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-nb\") pod \"20e53f85-d85c-466f-84b5-373d6d6af76b\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.672606 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-svc\") pod \"20e53f85-d85c-466f-84b5-373d6d6af76b\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.672632 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-config\") pod \"20e53f85-d85c-466f-84b5-373d6d6af76b\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.672654 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-swift-storage-0\") pod \"20e53f85-d85c-466f-84b5-373d6d6af76b\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.672690 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-sb\") pod \"20e53f85-d85c-466f-84b5-373d6d6af76b\" (UID: \"20e53f85-d85c-466f-84b5-373d6d6af76b\") " Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.681469 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20e53f85-d85c-466f-84b5-373d6d6af76b-kube-api-access-2cs68" (OuterVolumeSpecName: "kube-api-access-2cs68") pod "20e53f85-d85c-466f-84b5-373d6d6af76b" (UID: "20e53f85-d85c-466f-84b5-373d6d6af76b"). InnerVolumeSpecName "kube-api-access-2cs68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.746730 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "20e53f85-d85c-466f-84b5-373d6d6af76b" (UID: "20e53f85-d85c-466f-84b5-373d6d6af76b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.748420 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-config" (OuterVolumeSpecName: "config") pod "20e53f85-d85c-466f-84b5-373d6d6af76b" (UID: "20e53f85-d85c-466f-84b5-373d6d6af76b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.761945 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20e53f85-d85c-466f-84b5-373d6d6af76b" (UID: "20e53f85-d85c-466f-84b5-373d6d6af76b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.775919 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.775957 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.775982 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.775991 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cs68\" (UniqueName: \"kubernetes.io/projected/20e53f85-d85c-466f-84b5-373d6d6af76b-kube-api-access-2cs68\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.791462 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "20e53f85-d85c-466f-84b5-373d6d6af76b" (UID: "20e53f85-d85c-466f-84b5-373d6d6af76b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.793742 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-59f9c4fc59-kvs2g"] Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.812502 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20e53f85-d85c-466f-84b5-373d6d6af76b" (UID: "20e53f85-d85c-466f-84b5-373d6d6af76b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.878024 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:32 crc kubenswrapper[4810]: I1008 06:49:32.878056 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20e53f85-d85c-466f-84b5-373d6d6af76b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.096361 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.197954 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f98c8c4f6-259sf"] Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.218458 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c78787df7-xcmqq"] Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.232206 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5b4bcc558b-zx56q"] Oct 08 06:49:33 crc kubenswrapper[4810]: W1008 06:49:33.246602 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaabc3b24_1971_4326_aff0_12ddbf1f5ea1.slice/crio-71873762cf344d06269eb1a0804ef2572b9f9c16eda30852737b8d83a5a82a1e WatchSource:0}: Error finding container 71873762cf344d06269eb1a0804ef2572b9f9c16eda30852737b8d83a5a82a1e: Status 404 returned error can't find the container with id 71873762cf344d06269eb1a0804ef2572b9f9c16eda30852737b8d83a5a82a1e Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.539655 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" event={"ID":"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1","Type":"ContainerStarted","Data":"a1da74fa58f3eb079cecd485ee70af2de2336833d2f21c9da92f00fc742fa804"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.546318 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.546743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b55c5465-xfkrq" event={"ID":"20e53f85-d85c-466f-84b5-373d6d6af76b","Type":"ContainerDied","Data":"07d5ca0d0e67c7e578245d6c151d4f2346161a013685b2e9617411cf6d02999a"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.547578 4810 scope.go:117] "RemoveContainer" containerID="bf6d696953c6d7d34eb261bbfc2d0cec27de8b37a20463601d0cd4f3c9ec1703" Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.548922 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerStarted","Data":"c33f6599a9b7c96830053f6f56620e823df038dc79a79b964efc8109d7c6aff1"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.551183 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f98c8c4f6-259sf" event={"ID":"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34","Type":"ContainerStarted","Data":"7fc43758abdc8db60197440a965b398ff82c2f889ee8b4c7663bdec89c3e7617"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.551214 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f98c8c4f6-259sf" event={"ID":"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34","Type":"ContainerStarted","Data":"1c5071144e9209494241aaefd22f179cc359856f3f9f662a80a2574f79564e4a"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.553947 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" event={"ID":"791f9b40-309e-4266-9962-919281a31ffa","Type":"ContainerStarted","Data":"45b68c993c59b693c259ee3bbca5cedcb1d36e22d8d2bb51bfc0b245f89387f8"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.553990 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" event={"ID":"791f9b40-309e-4266-9962-919281a31ffa","Type":"ContainerStarted","Data":"c4f86d7de06ba0010017ba75fa49affc674ca8ee38a5c220c92c55b06940f7ed"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.556780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" event={"ID":"aabc3b24-1971-4326-aff0-12ddbf1f5ea1","Type":"ContainerStarted","Data":"71873762cf344d06269eb1a0804ef2572b9f9c16eda30852737b8d83a5a82a1e"} Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.733419 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b55c5465-xfkrq"] Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.744290 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b55c5465-xfkrq"] Oct 08 06:49:33 crc kubenswrapper[4810]: I1008 06:49:33.744665 4810 scope.go:117] "RemoveContainer" containerID="6501f4b4452bb73d358f200ef3d7c9bc4f681eb90be955d42bf54c0341695bbe" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.169717 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20e53f85-d85c-466f-84b5-373d6d6af76b" path="/var/lib/kubelet/pods/20e53f85-d85c-466f-84b5-373d6d6af76b/volumes" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.570861 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerStarted","Data":"f49a2aff835e211c0fb4f425b3e61294f7a786a0b04be605156e98801045442f"} Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.576319 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f98c8c4f6-259sf" event={"ID":"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34","Type":"ContainerStarted","Data":"df5060d62f2b8d65290b55dea47dd05f62641000a36b3f6ac8d6189c0244945a"} Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.576514 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.583041 4810 generic.go:334] "Generic (PLEG): container finished" podID="791f9b40-309e-4266-9962-919281a31ffa" containerID="45b68c993c59b693c259ee3bbca5cedcb1d36e22d8d2bb51bfc0b245f89387f8" exitCode=0 Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.584246 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" event={"ID":"791f9b40-309e-4266-9962-919281a31ffa","Type":"ContainerDied","Data":"45b68c993c59b693c259ee3bbca5cedcb1d36e22d8d2bb51bfc0b245f89387f8"} Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.599128 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f98c8c4f6-259sf" podStartSLOduration=3.599112773 podStartE2EDuration="3.599112773s" podCreationTimestamp="2025-10-08 06:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:34.593782525 +0000 UTC m=+1077.228222265" watchObservedRunningTime="2025-10-08 06:49:34.599112773 +0000 UTC m=+1077.233552513" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.971911 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6547ddf7cb-m6hp6"] Oct 08 06:49:34 crc kubenswrapper[4810]: E1008 06:49:34.972676 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerName="dnsmasq-dns" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.972695 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerName="dnsmasq-dns" Oct 08 06:49:34 crc kubenswrapper[4810]: E1008 06:49:34.972721 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerName="init" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.972729 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerName="init" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.972904 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="20e53f85-d85c-466f-84b5-373d6d6af76b" containerName="dnsmasq-dns" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.974337 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.986556 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.987529 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 08 06:49:34 crc kubenswrapper[4810]: I1008 06:49:34.988441 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6547ddf7cb-m6hp6"] Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.040808 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016c35ce-3072-485f-86b8-bbe927f78f81-logs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.040879 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-combined-ca-bundle\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.040928 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwhs4\" (UniqueName: \"kubernetes.io/projected/016c35ce-3072-485f-86b8-bbe927f78f81-kube-api-access-lwhs4\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.040973 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-public-tls-certs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.040997 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data-custom\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.041042 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-internal-tls-certs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.041061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.143617 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data-custom\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.143731 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-internal-tls-certs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.143756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.143918 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016c35ce-3072-485f-86b8-bbe927f78f81-logs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.143973 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-combined-ca-bundle\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.144093 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwhs4\" (UniqueName: \"kubernetes.io/projected/016c35ce-3072-485f-86b8-bbe927f78f81-kube-api-access-lwhs4\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.144136 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-public-tls-certs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.145201 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016c35ce-3072-485f-86b8-bbe927f78f81-logs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.148583 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-public-tls-certs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.148632 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-internal-tls-certs\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.149583 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data-custom\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.150072 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-combined-ca-bundle\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.150955 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.164022 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwhs4\" (UniqueName: \"kubernetes.io/projected/016c35ce-3072-485f-86b8-bbe927f78f81-kube-api-access-lwhs4\") pod \"barbican-api-6547ddf7cb-m6hp6\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.349347 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.634109 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" event={"ID":"791f9b40-309e-4266-9962-919281a31ffa","Type":"ContainerStarted","Data":"71222aa26baec746c9bdfce055269e0b6e48395ed1fd372b8dbed9e415120ba7"} Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.634776 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.636871 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" event={"ID":"aabc3b24-1971-4326-aff0-12ddbf1f5ea1","Type":"ContainerStarted","Data":"585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f"} Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.639196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" event={"ID":"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1","Type":"ContainerStarted","Data":"f651cc8c1eb5d6258b63b8e4bc2df041f992fce6e5d8e0d734458bee6b085f54"} Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.641314 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerStarted","Data":"46ade48a3d5d9bb318477732f52abee69b508cfdd65092bd87c95120f584c0c1"} Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.641397 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.659649 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" podStartSLOduration=4.6596221270000004 podStartE2EDuration="4.659622127s" podCreationTimestamp="2025-10-08 06:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:35.656204912 +0000 UTC m=+1078.290644662" watchObservedRunningTime="2025-10-08 06:49:35.659622127 +0000 UTC m=+1078.294061867" Oct 08 06:49:35 crc kubenswrapper[4810]: I1008 06:49:35.893739 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6547ddf7cb-m6hp6"] Oct 08 06:49:35 crc kubenswrapper[4810]: W1008 06:49:35.900027 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod016c35ce_3072_485f_86b8_bbe927f78f81.slice/crio-dde73594c1882d70f3f7fe51a75d2da5070bc756ffd409defa4bcc2e723d7a85 WatchSource:0}: Error finding container dde73594c1882d70f3f7fe51a75d2da5070bc756ffd409defa4bcc2e723d7a85: Status 404 returned error can't find the container with id dde73594c1882d70f3f7fe51a75d2da5070bc756ffd409defa4bcc2e723d7a85 Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.653300 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerStarted","Data":"f909f07fcb2d1d66e1a1e528a9f89448a8d5c84c645128310a2a1ee246eed7f4"} Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.660769 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6547ddf7cb-m6hp6" event={"ID":"016c35ce-3072-485f-86b8-bbe927f78f81","Type":"ContainerStarted","Data":"378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2"} Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.660832 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6547ddf7cb-m6hp6" event={"ID":"016c35ce-3072-485f-86b8-bbe927f78f81","Type":"ContainerStarted","Data":"d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b"} Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.660905 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.660929 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6547ddf7cb-m6hp6" event={"ID":"016c35ce-3072-485f-86b8-bbe927f78f81","Type":"ContainerStarted","Data":"dde73594c1882d70f3f7fe51a75d2da5070bc756ffd409defa4bcc2e723d7a85"} Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.660948 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.664343 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" event={"ID":"aabc3b24-1971-4326-aff0-12ddbf1f5ea1","Type":"ContainerStarted","Data":"833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b"} Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.666589 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" event={"ID":"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1","Type":"ContainerStarted","Data":"f1481a876ddf5f2a5996c05b5e44aa7cfa1cf846a5a9246337a1abce43f43d18"} Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.699001 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6547ddf7cb-m6hp6" podStartSLOduration=2.698977834 podStartE2EDuration="2.698977834s" podCreationTimestamp="2025-10-08 06:49:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:36.688462453 +0000 UTC m=+1079.322902193" watchObservedRunningTime="2025-10-08 06:49:36.698977834 +0000 UTC m=+1079.333417574" Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.714002 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" podStartSLOduration=3.467638034 podStartE2EDuration="5.713982851s" podCreationTimestamp="2025-10-08 06:49:31 +0000 UTC" firstStartedPulling="2025-10-08 06:49:32.806834294 +0000 UTC m=+1075.441274024" lastFinishedPulling="2025-10-08 06:49:35.053179091 +0000 UTC m=+1077.687618841" observedRunningTime="2025-10-08 06:49:36.71286593 +0000 UTC m=+1079.347305670" watchObservedRunningTime="2025-10-08 06:49:36.713982851 +0000 UTC m=+1079.348422591" Oct 08 06:49:36 crc kubenswrapper[4810]: I1008 06:49:36.753840 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" podStartSLOduration=3.938232841 podStartE2EDuration="5.753818416s" podCreationTimestamp="2025-10-08 06:49:31 +0000 UTC" firstStartedPulling="2025-10-08 06:49:33.249019193 +0000 UTC m=+1075.883458923" lastFinishedPulling="2025-10-08 06:49:35.064604758 +0000 UTC m=+1077.699044498" observedRunningTime="2025-10-08 06:49:36.748283213 +0000 UTC m=+1079.382722953" watchObservedRunningTime="2025-10-08 06:49:36.753818416 +0000 UTC m=+1079.388258156" Oct 08 06:49:38 crc kubenswrapper[4810]: I1008 06:49:38.595530 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:49:39 crc kubenswrapper[4810]: I1008 06:49:39.722109 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerStarted","Data":"1a2b0b1aa72a97a60b12d9698daaf81f42dba0f5615ae1e65c3ade779bf70b25"} Oct 08 06:49:39 crc kubenswrapper[4810]: I1008 06:49:39.724204 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 06:49:39 crc kubenswrapper[4810]: I1008 06:49:39.759370 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.20325751 podStartE2EDuration="8.759348587s" podCreationTimestamp="2025-10-08 06:49:31 +0000 UTC" firstStartedPulling="2025-10-08 06:49:33.097920691 +0000 UTC m=+1075.732360431" lastFinishedPulling="2025-10-08 06:49:38.654011768 +0000 UTC m=+1081.288451508" observedRunningTime="2025-10-08 06:49:39.751749176 +0000 UTC m=+1082.386188926" watchObservedRunningTime="2025-10-08 06:49:39.759348587 +0000 UTC m=+1082.393788327" Oct 08 06:49:40 crc kubenswrapper[4810]: I1008 06:49:40.739162 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wqx77" event={"ID":"cb95c243-0c64-4937-a86e-34f348023513","Type":"ContainerStarted","Data":"e3d36cfbaa8984ab3a07deadd18a85cb569ec7b4cff4e318a3ad985e94a1df6a"} Oct 08 06:49:40 crc kubenswrapper[4810]: I1008 06:49:40.763376 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-wqx77" podStartSLOduration=2.812893906 podStartE2EDuration="41.763360664s" podCreationTimestamp="2025-10-08 06:48:59 +0000 UTC" firstStartedPulling="2025-10-08 06:49:00.566106323 +0000 UTC m=+1043.200546063" lastFinishedPulling="2025-10-08 06:49:39.516573081 +0000 UTC m=+1082.151012821" observedRunningTime="2025-10-08 06:49:40.75887689 +0000 UTC m=+1083.393316630" watchObservedRunningTime="2025-10-08 06:49:40.763360664 +0000 UTC m=+1083.397800404" Oct 08 06:49:42 crc kubenswrapper[4810]: I1008 06:49:42.379305 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:49:42 crc kubenswrapper[4810]: I1008 06:49:42.442840 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dc68bd5-mpslw"] Oct 08 06:49:42 crc kubenswrapper[4810]: I1008 06:49:42.444030 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" podUID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerName="dnsmasq-dns" containerID="cri-o://cafd24014584c1d4c9cf578ffee08a99f18df72ab6bfd6c4794466651cfa79fe" gracePeriod=10 Oct 08 06:49:42 crc kubenswrapper[4810]: I1008 06:49:42.801102 4810 generic.go:334] "Generic (PLEG): container finished" podID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerID="cafd24014584c1d4c9cf578ffee08a99f18df72ab6bfd6c4794466651cfa79fe" exitCode=0 Oct 08 06:49:42 crc kubenswrapper[4810]: I1008 06:49:42.801153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" event={"ID":"9d5fc753-8316-4e77-8e4a-c0ab288b6589","Type":"ContainerDied","Data":"cafd24014584c1d4c9cf578ffee08a99f18df72ab6bfd6c4794466651cfa79fe"} Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.190599 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.293754 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 06:49:43 crc kubenswrapper[4810]: E1008 06:49:43.294279 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerName="init" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.294299 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerName="init" Oct 08 06:49:43 crc kubenswrapper[4810]: E1008 06:49:43.294319 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerName="dnsmasq-dns" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.294327 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerName="dnsmasq-dns" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.294568 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" containerName="dnsmasq-dns" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.295267 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.300017 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-44qwd" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.300213 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.301665 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.305321 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.347733 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-nb\") pod \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.347798 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-svc\") pod \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.347869 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-swift-storage-0\") pod \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.347905 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gbl6\" (UniqueName: \"kubernetes.io/projected/9d5fc753-8316-4e77-8e4a-c0ab288b6589-kube-api-access-2gbl6\") pod \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.348012 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-config\") pod \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.348102 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-sb\") pod \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\" (UID: \"9d5fc753-8316-4e77-8e4a-c0ab288b6589\") " Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.368602 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d5fc753-8316-4e77-8e4a-c0ab288b6589-kube-api-access-2gbl6" (OuterVolumeSpecName: "kube-api-access-2gbl6") pod "9d5fc753-8316-4e77-8e4a-c0ab288b6589" (UID: "9d5fc753-8316-4e77-8e4a-c0ab288b6589"). InnerVolumeSpecName "kube-api-access-2gbl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.422363 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-config" (OuterVolumeSpecName: "config") pod "9d5fc753-8316-4e77-8e4a-c0ab288b6589" (UID: "9d5fc753-8316-4e77-8e4a-c0ab288b6589"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.436511 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d5fc753-8316-4e77-8e4a-c0ab288b6589" (UID: "9d5fc753-8316-4e77-8e4a-c0ab288b6589"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.449408 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7655f5c6df-qtz48"] Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.451226 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.452101 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxbrr\" (UniqueName: \"kubernetes.io/projected/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-kube-api-access-fxbrr\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.452174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config-secret\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.452207 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.452237 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.452653 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.452674 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gbl6\" (UniqueName: \"kubernetes.io/projected/9d5fc753-8316-4e77-8e4a-c0ab288b6589-kube-api-access-2gbl6\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.452712 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.456349 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.456404 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.456623 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.468243 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9d5fc753-8316-4e77-8e4a-c0ab288b6589" (UID: "9d5fc753-8316-4e77-8e4a-c0ab288b6589"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.489594 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d5fc753-8316-4e77-8e4a-c0ab288b6589" (UID: "9d5fc753-8316-4e77-8e4a-c0ab288b6589"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.490467 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7655f5c6df-qtz48"] Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.504923 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9d5fc753-8316-4e77-8e4a-c0ab288b6589" (UID: "9d5fc753-8316-4e77-8e4a-c0ab288b6589"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-log-httpd\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561484 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-combined-ca-bundle\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561548 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-config-data\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561599 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxbrr\" (UniqueName: \"kubernetes.io/projected/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-kube-api-access-fxbrr\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561648 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnctt\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-kube-api-access-xnctt\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561680 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config-secret\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561849 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.561941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-internal-tls-certs\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.562015 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-public-tls-certs\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.562058 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.562100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-etc-swift\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.562334 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-run-httpd\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.562495 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.562519 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.562531 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d5fc753-8316-4e77-8e4a-c0ab288b6589-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.564753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.568282 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config-secret\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.568911 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.582612 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxbrr\" (UniqueName: \"kubernetes.io/projected/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-kube-api-access-fxbrr\") pod \"openstackclient\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.613439 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.667677 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-config-data\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.667809 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnctt\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-kube-api-access-xnctt\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.667887 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-internal-tls-certs\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.667915 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-public-tls-certs\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.667941 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-etc-swift\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.668037 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-run-httpd\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.668079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-log-httpd\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.668124 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-combined-ca-bundle\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.668756 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-run-httpd\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.669049 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-log-httpd\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.675653 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-combined-ca-bundle\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.676825 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-config-data\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.700585 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-public-tls-certs\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.701058 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-internal-tls-certs\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.704208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-etc-swift\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.734724 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnctt\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-kube-api-access-xnctt\") pod \"swift-proxy-7655f5c6df-qtz48\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.815225 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" event={"ID":"9d5fc753-8316-4e77-8e4a-c0ab288b6589","Type":"ContainerDied","Data":"0f50f021b43579d636062d4ec7e244b3fcf631eb1d2bf79db6e4ec97b62f2eca"} Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.815292 4810 scope.go:117] "RemoveContainer" containerID="cafd24014584c1d4c9cf578ffee08a99f18df72ab6bfd6c4794466651cfa79fe" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.815503 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dc68bd5-mpslw" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.836151 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.879171 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dc68bd5-mpslw"] Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.886735 4810 scope.go:117] "RemoveContainer" containerID="dc33a789cf086b565ca26150f8a066de6c4e57ed631aabf50ab45142b77970ef" Oct 08 06:49:43 crc kubenswrapper[4810]: I1008 06:49:43.887718 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dc68bd5-mpslw"] Oct 08 06:49:44 crc kubenswrapper[4810]: I1008 06:49:44.089189 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d5fc753-8316-4e77-8e4a-c0ab288b6589" path="/var/lib/kubelet/pods/9d5fc753-8316-4e77-8e4a-c0ab288b6589/volumes" Oct 08 06:49:44 crc kubenswrapper[4810]: I1008 06:49:44.131800 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 06:49:44 crc kubenswrapper[4810]: W1008 06:49:44.142783 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31bfa5d9_dc8e_495d_983a_a17bd5a9b11f.slice/crio-aa007bb9ce821a31e1c00cbcd246de0fab679caaccd1ff9d32c5a4c23af38f99 WatchSource:0}: Error finding container aa007bb9ce821a31e1c00cbcd246de0fab679caaccd1ff9d32c5a4c23af38f99: Status 404 returned error can't find the container with id aa007bb9ce821a31e1c00cbcd246de0fab679caaccd1ff9d32c5a4c23af38f99 Oct 08 06:49:44 crc kubenswrapper[4810]: I1008 06:49:44.584702 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:44 crc kubenswrapper[4810]: I1008 06:49:44.830873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f","Type":"ContainerStarted","Data":"aa007bb9ce821a31e1c00cbcd246de0fab679caaccd1ff9d32c5a4c23af38f99"} Oct 08 06:49:44 crc kubenswrapper[4810]: I1008 06:49:44.849471 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7655f5c6df-qtz48"] Oct 08 06:49:45 crc kubenswrapper[4810]: I1008 06:49:45.005621 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:45 crc kubenswrapper[4810]: I1008 06:49:45.860259 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7655f5c6df-qtz48" event={"ID":"06237db0-424d-4bd7-ae4f-3f7af9cc4f92","Type":"ContainerStarted","Data":"497c584cc6367d012ea7e0e3c2b067d6f076bf9d675830debc76b79c7a56c2ca"} Oct 08 06:49:45 crc kubenswrapper[4810]: I1008 06:49:45.860557 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7655f5c6df-qtz48" event={"ID":"06237db0-424d-4bd7-ae4f-3f7af9cc4f92","Type":"ContainerStarted","Data":"67eb5ec810014b1c8f7679774ea7850e6282168a5aaf2c5cf4b56dfb7bddaec8"} Oct 08 06:49:45 crc kubenswrapper[4810]: I1008 06:49:45.860574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7655f5c6df-qtz48" event={"ID":"06237db0-424d-4bd7-ae4f-3f7af9cc4f92","Type":"ContainerStarted","Data":"7657cdad3606aa5cab96c05e2cbd935fffbbc42a1a0d867410bb359ed6fc9406"} Oct 08 06:49:45 crc kubenswrapper[4810]: I1008 06:49:45.861810 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:45 crc kubenswrapper[4810]: I1008 06:49:45.861899 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:45 crc kubenswrapper[4810]: I1008 06:49:45.897822 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7655f5c6df-qtz48" podStartSLOduration=2.897800172 podStartE2EDuration="2.897800172s" podCreationTimestamp="2025-10-08 06:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:45.888030851 +0000 UTC m=+1088.522470611" watchObservedRunningTime="2025-10-08 06:49:45.897800172 +0000 UTC m=+1088.532239912" Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.490612 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.490995 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-central-agent" containerID="cri-o://f49a2aff835e211c0fb4f425b3e61294f7a786a0b04be605156e98801045442f" gracePeriod=30 Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.491070 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="proxy-httpd" containerID="cri-o://1a2b0b1aa72a97a60b12d9698daaf81f42dba0f5615ae1e65c3ade779bf70b25" gracePeriod=30 Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.491117 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-notification-agent" containerID="cri-o://46ade48a3d5d9bb318477732f52abee69b508cfdd65092bd87c95120f584c0c1" gracePeriod=30 Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.491338 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="sg-core" containerID="cri-o://f909f07fcb2d1d66e1a1e528a9f89448a8d5c84c645128310a2a1ee246eed7f4" gracePeriod=30 Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.875800 4810 generic.go:334] "Generic (PLEG): container finished" podID="cb95c243-0c64-4937-a86e-34f348023513" containerID="e3d36cfbaa8984ab3a07deadd18a85cb569ec7b4cff4e318a3ad985e94a1df6a" exitCode=0 Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.875888 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wqx77" event={"ID":"cb95c243-0c64-4937-a86e-34f348023513","Type":"ContainerDied","Data":"e3d36cfbaa8984ab3a07deadd18a85cb569ec7b4cff4e318a3ad985e94a1df6a"} Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.890389 4810 generic.go:334] "Generic (PLEG): container finished" podID="da3b3c98-166f-433f-8e3b-38125745f14f" containerID="1a2b0b1aa72a97a60b12d9698daaf81f42dba0f5615ae1e65c3ade779bf70b25" exitCode=0 Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.890664 4810 generic.go:334] "Generic (PLEG): container finished" podID="da3b3c98-166f-433f-8e3b-38125745f14f" containerID="f909f07fcb2d1d66e1a1e528a9f89448a8d5c84c645128310a2a1ee246eed7f4" exitCode=2 Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.890480 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerDied","Data":"1a2b0b1aa72a97a60b12d9698daaf81f42dba0f5615ae1e65c3ade779bf70b25"} Oct 08 06:49:46 crc kubenswrapper[4810]: I1008 06:49:46.891161 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerDied","Data":"f909f07fcb2d1d66e1a1e528a9f89448a8d5c84c645128310a2a1ee246eed7f4"} Oct 08 06:49:47 crc kubenswrapper[4810]: I1008 06:49:47.908195 4810 generic.go:334] "Generic (PLEG): container finished" podID="da3b3c98-166f-433f-8e3b-38125745f14f" containerID="f49a2aff835e211c0fb4f425b3e61294f7a786a0b04be605156e98801045442f" exitCode=0 Oct 08 06:49:47 crc kubenswrapper[4810]: I1008 06:49:47.908265 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerDied","Data":"f49a2aff835e211c0fb4f425b3e61294f7a786a0b04be605156e98801045442f"} Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.201333 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.377578 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wqx77" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.494638 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-db-sync-config-data\") pod \"cb95c243-0c64-4937-a86e-34f348023513\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.494819 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-combined-ca-bundle\") pod \"cb95c243-0c64-4937-a86e-34f348023513\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.494862 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-scripts\") pod \"cb95c243-0c64-4937-a86e-34f348023513\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.494904 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb95c243-0c64-4937-a86e-34f348023513-etc-machine-id\") pod \"cb95c243-0c64-4937-a86e-34f348023513\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.495048 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbpv4\" (UniqueName: \"kubernetes.io/projected/cb95c243-0c64-4937-a86e-34f348023513-kube-api-access-mbpv4\") pod \"cb95c243-0c64-4937-a86e-34f348023513\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.495109 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-config-data\") pod \"cb95c243-0c64-4937-a86e-34f348023513\" (UID: \"cb95c243-0c64-4937-a86e-34f348023513\") " Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.495846 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb95c243-0c64-4937-a86e-34f348023513-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb95c243-0c64-4937-a86e-34f348023513" (UID: "cb95c243-0c64-4937-a86e-34f348023513"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.501924 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-scripts" (OuterVolumeSpecName: "scripts") pod "cb95c243-0c64-4937-a86e-34f348023513" (UID: "cb95c243-0c64-4937-a86e-34f348023513"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.518189 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cb95c243-0c64-4937-a86e-34f348023513" (UID: "cb95c243-0c64-4937-a86e-34f348023513"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.520426 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb95c243-0c64-4937-a86e-34f348023513-kube-api-access-mbpv4" (OuterVolumeSpecName: "kube-api-access-mbpv4") pod "cb95c243-0c64-4937-a86e-34f348023513" (UID: "cb95c243-0c64-4937-a86e-34f348023513"). InnerVolumeSpecName "kube-api-access-mbpv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.552861 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-config-data" (OuterVolumeSpecName: "config-data") pod "cb95c243-0c64-4937-a86e-34f348023513" (UID: "cb95c243-0c64-4937-a86e-34f348023513"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.555531 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.560672 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb95c243-0c64-4937-a86e-34f348023513" (UID: "cb95c243-0c64-4937-a86e-34f348023513"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.651358 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f98c8c4f6-259sf"] Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.651652 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f98c8c4f6-259sf" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api-log" containerID="cri-o://7fc43758abdc8db60197440a965b398ff82c2f889ee8b4c7663bdec89c3e7617" gracePeriod=30 Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.651938 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f98c8c4f6-259sf" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api" containerID="cri-o://df5060d62f2b8d65290b55dea47dd05f62641000a36b3f6ac8d6189c0244945a" gracePeriod=30 Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.598951 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.673149 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.673234 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.673317 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb95c243-0c64-4937-a86e-34f348023513-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.673434 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbpv4\" (UniqueName: \"kubernetes.io/projected/cb95c243-0c64-4937-a86e-34f348023513-kube-api-access-mbpv4\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.673537 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb95c243-0c64-4937-a86e-34f348023513-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.936216 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wqx77" event={"ID":"cb95c243-0c64-4937-a86e-34f348023513","Type":"ContainerDied","Data":"c70e42e3f471ef7f555620f357e1f14834d3321f897ce4e89df12bf910317d62"} Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.936263 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c70e42e3f471ef7f555620f357e1f14834d3321f897ce4e89df12bf910317d62" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.936270 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wqx77" Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.941062 4810 generic.go:334] "Generic (PLEG): container finished" podID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerID="7fc43758abdc8db60197440a965b398ff82c2f889ee8b4c7663bdec89c3e7617" exitCode=143 Oct 08 06:49:48 crc kubenswrapper[4810]: I1008 06:49:48.942096 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f98c8c4f6-259sf" event={"ID":"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34","Type":"ContainerDied","Data":"7fc43758abdc8db60197440a965b398ff82c2f889ee8b4c7663bdec89c3e7617"} Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.298840 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84bd785c49-86blx"] Oct 08 06:49:49 crc kubenswrapper[4810]: E1008 06:49:49.301230 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb95c243-0c64-4937-a86e-34f348023513" containerName="cinder-db-sync" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.301252 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb95c243-0c64-4937-a86e-34f348023513" containerName="cinder-db-sync" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.301449 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb95c243-0c64-4937-a86e-34f348023513" containerName="cinder-db-sync" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.302511 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.319017 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bd785c49-86blx"] Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.387390 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-svc\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.387573 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcf4t\" (UniqueName: \"kubernetes.io/projected/850b635b-9e46-4f19-b573-9dcde8ec4d24-kube-api-access-rcf4t\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.387613 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-swift-storage-0\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.387638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-config\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.387952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-nb\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.388012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-sb\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.431765 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.431819 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.439998 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.441768 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.445860 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9xsgq" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.446128 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.446458 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.446640 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.458339 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.492115 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-config\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.492194 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.492495 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9mth\" (UniqueName: \"kubernetes.io/projected/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-kube-api-access-l9mth\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.492648 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.492730 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-nb\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.492755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-sb\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.492830 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.493143 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-svc\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.493237 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-scripts\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.493335 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcf4t\" (UniqueName: \"kubernetes.io/projected/850b635b-9e46-4f19-b573-9dcde8ec4d24-kube-api-access-rcf4t\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.493386 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.493434 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-swift-storage-0\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.496533 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-nb\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.496625 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-svc\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.496653 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-sb\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.496945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-swift-storage-0\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.497374 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-config\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.526354 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcf4t\" (UniqueName: \"kubernetes.io/projected/850b635b-9e46-4f19-b573-9dcde8ec4d24-kube-api-access-rcf4t\") pod \"dnsmasq-dns-84bd785c49-86blx\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.595779 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.595900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-scripts\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.596527 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.596609 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.596661 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9mth\" (UniqueName: \"kubernetes.io/projected/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-kube-api-access-l9mth\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.596790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.596815 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.602653 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.602714 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-scripts\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.606999 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.616593 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.623531 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9mth\" (UniqueName: \"kubernetes.io/projected/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-kube-api-access-l9mth\") pod \"cinder-scheduler-0\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.636767 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.646064 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.648524 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.652645 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.661433 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.698484 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aca4e1b-044e-4f63-8af1-51e77d5acce4-logs\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.698754 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-scripts\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.698813 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sscc7\" (UniqueName: \"kubernetes.io/projected/5aca4e1b-044e-4f63-8af1-51e77d5acce4-kube-api-access-sscc7\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.698835 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.698889 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.698911 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data-custom\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.699009 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5aca4e1b-044e-4f63-8af1-51e77d5acce4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.778262 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.800325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.800373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data-custom\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.800445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5aca4e1b-044e-4f63-8af1-51e77d5acce4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.800471 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aca4e1b-044e-4f63-8af1-51e77d5acce4-logs\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.800499 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-scripts\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.800549 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sscc7\" (UniqueName: \"kubernetes.io/projected/5aca4e1b-044e-4f63-8af1-51e77d5acce4-kube-api-access-sscc7\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.800571 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.801484 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5aca4e1b-044e-4f63-8af1-51e77d5acce4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.806134 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.807812 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aca4e1b-044e-4f63-8af1-51e77d5acce4-logs\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.808411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.809405 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-scripts\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.809856 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data-custom\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.826617 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sscc7\" (UniqueName: \"kubernetes.io/projected/5aca4e1b-044e-4f63-8af1-51e77d5acce4-kube-api-access-sscc7\") pod \"cinder-api-0\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " pod="openstack/cinder-api-0" Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.971523 4810 generic.go:334] "Generic (PLEG): container finished" podID="da3b3c98-166f-433f-8e3b-38125745f14f" containerID="46ade48a3d5d9bb318477732f52abee69b508cfdd65092bd87c95120f584c0c1" exitCode=0 Oct 08 06:49:49 crc kubenswrapper[4810]: I1008 06:49:49.971567 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerDied","Data":"46ade48a3d5d9bb318477732f52abee69b508cfdd65092bd87c95120f584c0c1"} Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.087298 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.187727 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bd785c49-86blx"] Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.357679 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.518110 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.618102 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-run-httpd\") pod \"da3b3c98-166f-433f-8e3b-38125745f14f\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.618154 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-log-httpd\") pod \"da3b3c98-166f-433f-8e3b-38125745f14f\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.618258 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql526\" (UniqueName: \"kubernetes.io/projected/da3b3c98-166f-433f-8e3b-38125745f14f-kube-api-access-ql526\") pod \"da3b3c98-166f-433f-8e3b-38125745f14f\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.618311 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-scripts\") pod \"da3b3c98-166f-433f-8e3b-38125745f14f\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.618369 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-sg-core-conf-yaml\") pod \"da3b3c98-166f-433f-8e3b-38125745f14f\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.618708 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "da3b3c98-166f-433f-8e3b-38125745f14f" (UID: "da3b3c98-166f-433f-8e3b-38125745f14f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.619611 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-combined-ca-bundle\") pod \"da3b3c98-166f-433f-8e3b-38125745f14f\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.619674 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-config-data\") pod \"da3b3c98-166f-433f-8e3b-38125745f14f\" (UID: \"da3b3c98-166f-433f-8e3b-38125745f14f\") " Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.620317 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.621906 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "da3b3c98-166f-433f-8e3b-38125745f14f" (UID: "da3b3c98-166f-433f-8e3b-38125745f14f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.625167 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da3b3c98-166f-433f-8e3b-38125745f14f-kube-api-access-ql526" (OuterVolumeSpecName: "kube-api-access-ql526") pod "da3b3c98-166f-433f-8e3b-38125745f14f" (UID: "da3b3c98-166f-433f-8e3b-38125745f14f"). InnerVolumeSpecName "kube-api-access-ql526". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.631107 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-scripts" (OuterVolumeSpecName: "scripts") pod "da3b3c98-166f-433f-8e3b-38125745f14f" (UID: "da3b3c98-166f-433f-8e3b-38125745f14f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.667425 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.673556 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "da3b3c98-166f-433f-8e3b-38125745f14f" (UID: "da3b3c98-166f-433f-8e3b-38125745f14f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:50 crc kubenswrapper[4810]: W1008 06:49:50.678902 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5aca4e1b_044e_4f63_8af1_51e77d5acce4.slice/crio-e2c3c0f04cb614bafbd025ca6199adad6c4269b0172dd2560fafb504617d8728 WatchSource:0}: Error finding container e2c3c0f04cb614bafbd025ca6199adad6c4269b0172dd2560fafb504617d8728: Status 404 returned error can't find the container with id e2c3c0f04cb614bafbd025ca6199adad6c4269b0172dd2560fafb504617d8728 Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.721881 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.721912 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da3b3c98-166f-433f-8e3b-38125745f14f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.721923 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql526\" (UniqueName: \"kubernetes.io/projected/da3b3c98-166f-433f-8e3b-38125745f14f-kube-api-access-ql526\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.721933 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.739421 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da3b3c98-166f-433f-8e3b-38125745f14f" (UID: "da3b3c98-166f-433f-8e3b-38125745f14f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.772588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-config-data" (OuterVolumeSpecName: "config-data") pod "da3b3c98-166f-433f-8e3b-38125745f14f" (UID: "da3b3c98-166f-433f-8e3b-38125745f14f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.824361 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.824402 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da3b3c98-166f-433f-8e3b-38125745f14f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.983481 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa","Type":"ContainerStarted","Data":"42e8ce13ce2d8c5e543a4649dc3a29ea43b116d6e1f9378e2482b2b0e9c98bd9"} Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.989250 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da3b3c98-166f-433f-8e3b-38125745f14f","Type":"ContainerDied","Data":"c33f6599a9b7c96830053f6f56620e823df038dc79a79b964efc8109d7c6aff1"} Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.989336 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:50 crc kubenswrapper[4810]: I1008 06:49:50.989686 4810 scope.go:117] "RemoveContainer" containerID="1a2b0b1aa72a97a60b12d9698daaf81f42dba0f5615ae1e65c3ade779bf70b25" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.003391 4810 generic.go:334] "Generic (PLEG): container finished" podID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerID="193dc08e9cf522234431a6f7ee7bb9230f9702c5ab78ea227592b5701064d56e" exitCode=0 Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.003467 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bd785c49-86blx" event={"ID":"850b635b-9e46-4f19-b573-9dcde8ec4d24","Type":"ContainerDied","Data":"193dc08e9cf522234431a6f7ee7bb9230f9702c5ab78ea227592b5701064d56e"} Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.003511 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bd785c49-86blx" event={"ID":"850b635b-9e46-4f19-b573-9dcde8ec4d24","Type":"ContainerStarted","Data":"7c7ad6b73903a4a9805471de20f27c4c392c49ef8d2c90254a13bd7794140719"} Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.009901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5aca4e1b-044e-4f63-8af1-51e77d5acce4","Type":"ContainerStarted","Data":"e2c3c0f04cb614bafbd025ca6199adad6c4269b0172dd2560fafb504617d8728"} Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.026632 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.035151 4810 scope.go:117] "RemoveContainer" containerID="f909f07fcb2d1d66e1a1e528a9f89448a8d5c84c645128310a2a1ee246eed7f4" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.039498 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.123502 4810 scope.go:117] "RemoveContainer" containerID="46ade48a3d5d9bb318477732f52abee69b508cfdd65092bd87c95120f584c0c1" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.180154 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:51 crc kubenswrapper[4810]: E1008 06:49:51.180742 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="sg-core" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.180763 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="sg-core" Oct 08 06:49:51 crc kubenswrapper[4810]: E1008 06:49:51.180786 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="proxy-httpd" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.180792 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="proxy-httpd" Oct 08 06:49:51 crc kubenswrapper[4810]: E1008 06:49:51.180823 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-notification-agent" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.180830 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-notification-agent" Oct 08 06:49:51 crc kubenswrapper[4810]: E1008 06:49:51.180848 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-central-agent" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.180854 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-central-agent" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.181108 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="proxy-httpd" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.181126 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-notification-agent" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.181136 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="sg-core" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.181146 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" containerName="ceilometer-central-agent" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.183260 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.193094 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.194238 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.200175 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.213019 4810 scope.go:117] "RemoveContainer" containerID="f49a2aff835e211c0fb4f425b3e61294f7a786a0b04be605156e98801045442f" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.340533 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.340937 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-run-httpd\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.341009 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h5r5\" (UniqueName: \"kubernetes.io/projected/2c7f6b7e-7579-4441-9ec3-6a421478c21c-kube-api-access-2h5r5\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.341096 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-config-data\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.341141 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-log-httpd\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.341167 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-scripts\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.341189 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.445047 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-config-data\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.445123 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-log-httpd\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.445145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-scripts\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.445161 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.445241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.445267 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-run-httpd\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.445302 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h5r5\" (UniqueName: \"kubernetes.io/projected/2c7f6b7e-7579-4441-9ec3-6a421478c21c-kube-api-access-2h5r5\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.446290 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-run-httpd\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.446314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-log-httpd\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.450935 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-config-data\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.452045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.452077 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.452232 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-scripts\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.469067 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h5r5\" (UniqueName: \"kubernetes.io/projected/2c7f6b7e-7579-4441-9ec3-6a421478c21c-kube-api-access-2h5r5\") pod \"ceilometer-0\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " pod="openstack/ceilometer-0" Oct 08 06:49:51 crc kubenswrapper[4810]: I1008 06:49:51.516808 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.042746 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bd785c49-86blx" event={"ID":"850b635b-9e46-4f19-b573-9dcde8ec4d24","Type":"ContainerStarted","Data":"849c967153d2f165920e9dfd4ff8e9b6bd05ad3f465bda0698dd4e80edcf02a0"} Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.043867 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.049511 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5aca4e1b-044e-4f63-8af1-51e77d5acce4","Type":"ContainerStarted","Data":"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972"} Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.054189 4810 generic.go:334] "Generic (PLEG): container finished" podID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerID="df5060d62f2b8d65290b55dea47dd05f62641000a36b3f6ac8d6189c0244945a" exitCode=0 Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.054256 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f98c8c4f6-259sf" event={"ID":"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34","Type":"ContainerDied","Data":"df5060d62f2b8d65290b55dea47dd05f62641000a36b3f6ac8d6189c0244945a"} Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.075817 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84bd785c49-86blx" podStartSLOduration=3.075786485 podStartE2EDuration="3.075786485s" podCreationTimestamp="2025-10-08 06:49:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:52.067661419 +0000 UTC m=+1094.702101159" watchObservedRunningTime="2025-10-08 06:49:52.075786485 +0000 UTC m=+1094.710226235" Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.097591 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da3b3c98-166f-433f-8e3b-38125745f14f" path="/var/lib/kubelet/pods/da3b3c98-166f-433f-8e3b-38125745f14f/volumes" Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.397467 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f98c8c4f6-259sf" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.397858 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f98c8c4f6-259sf" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Oct 08 06:49:52 crc kubenswrapper[4810]: I1008 06:49:52.415719 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:49:53 crc kubenswrapper[4810]: I1008 06:49:53.237437 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:53 crc kubenswrapper[4810]: I1008 06:49:53.846356 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:53 crc kubenswrapper[4810]: I1008 06:49:53.847247 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.394722 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f98c8c4f6-259sf" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.394744 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f98c8c4f6-259sf" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.730863 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.887536 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.897820 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data\") pod \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.897881 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-logs\") pod \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.898000 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp465\" (UniqueName: \"kubernetes.io/projected/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-kube-api-access-vp465\") pod \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.898068 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-combined-ca-bundle\") pod \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.898172 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data-custom\") pod \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\" (UID: \"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34\") " Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.900295 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-logs" (OuterVolumeSpecName: "logs") pod "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" (UID: "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.905264 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-kube-api-access-vp465" (OuterVolumeSpecName: "kube-api-access-vp465") pod "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" (UID: "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34"). InnerVolumeSpecName "kube-api-access-vp465". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.907851 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" (UID: "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.955132 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" (UID: "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:57 crc kubenswrapper[4810]: I1008 06:49:57.991258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data" (OuterVolumeSpecName: "config-data") pod "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" (UID: "a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.000541 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.000577 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.000587 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp465\" (UniqueName: \"kubernetes.io/projected/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-kube-api-access-vp465\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.000596 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.000606 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.170808 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f","Type":"ContainerStarted","Data":"a1c4944dcfd55a27b1f280aa67cbd8b54e526e0b2822d8a76c6d1e57a1032b88"} Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.187990 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f98c8c4f6-259sf" event={"ID":"a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34","Type":"ContainerDied","Data":"1c5071144e9209494241aaefd22f179cc359856f3f9f662a80a2574f79564e4a"} Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.188055 4810 scope.go:117] "RemoveContainer" containerID="df5060d62f2b8d65290b55dea47dd05f62641000a36b3f6ac8d6189c0244945a" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.188303 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f98c8c4f6-259sf" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.219019 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.221669 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.756541817 podStartE2EDuration="15.221651875s" podCreationTimestamp="2025-10-08 06:49:43 +0000 UTC" firstStartedPulling="2025-10-08 06:49:44.146208613 +0000 UTC m=+1086.780648353" lastFinishedPulling="2025-10-08 06:49:57.611318651 +0000 UTC m=+1100.245758411" observedRunningTime="2025-10-08 06:49:58.197408172 +0000 UTC m=+1100.831847912" watchObservedRunningTime="2025-10-08 06:49:58.221651875 +0000 UTC m=+1100.856091615" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.236411 4810 scope.go:117] "RemoveContainer" containerID="7fc43758abdc8db60197440a965b398ff82c2f889ee8b4c7663bdec89c3e7617" Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.239756 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f98c8c4f6-259sf"] Oct 08 06:49:58 crc kubenswrapper[4810]: I1008 06:49:58.250581 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f98c8c4f6-259sf"] Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.199824 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerStarted","Data":"7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9"} Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.200546 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerStarted","Data":"9638549ecf6d13608addf6517e520bc46d95ec9ea9d2fa3a5070d815251ebd53"} Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.204543 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa","Type":"ContainerStarted","Data":"362be8af4c64a782d1ca3bc7cbfcd409e92cddc7ff4c65e3f4e36cb4913f9d77"} Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.204605 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa","Type":"ContainerStarted","Data":"def11983963f3bac578eb6868e32992fb949f4d82a70c0cdeab56ce2cc363305"} Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.207877 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5aca4e1b-044e-4f63-8af1-51e77d5acce4","Type":"ContainerStarted","Data":"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55"} Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.207977 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api-log" containerID="cri-o://c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972" gracePeriod=30 Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.207998 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.208086 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api" containerID="cri-o://2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55" gracePeriod=30 Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.234661 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.017630732 podStartE2EDuration="10.234642172s" podCreationTimestamp="2025-10-08 06:49:49 +0000 UTC" firstStartedPulling="2025-10-08 06:49:50.387209595 +0000 UTC m=+1093.021649335" lastFinishedPulling="2025-10-08 06:49:57.604221035 +0000 UTC m=+1100.238660775" observedRunningTime="2025-10-08 06:49:59.224413807 +0000 UTC m=+1101.858853547" watchObservedRunningTime="2025-10-08 06:49:59.234642172 +0000 UTC m=+1101.869081912" Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.257234 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=10.257205218 podStartE2EDuration="10.257205218s" podCreationTimestamp="2025-10-08 06:49:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:49:59.245305698 +0000 UTC m=+1101.879745438" watchObservedRunningTime="2025-10-08 06:49:59.257205218 +0000 UTC m=+1101.891644958" Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.639090 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.734805 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c78787df7-xcmqq"] Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.735056 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" podUID="791f9b40-309e-4266-9962-919281a31ffa" containerName="dnsmasq-dns" containerID="cri-o://71222aa26baec746c9bdfce055269e0b6e48395ed1fd372b8dbed9e415120ba7" gracePeriod=10 Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.778744 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 06:49:59 crc kubenswrapper[4810]: I1008 06:49:59.955485 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.087457 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" path="/var/lib/kubelet/pods/a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34/volumes" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149075 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sscc7\" (UniqueName: \"kubernetes.io/projected/5aca4e1b-044e-4f63-8af1-51e77d5acce4-kube-api-access-sscc7\") pod \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149124 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data-custom\") pod \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149195 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5aca4e1b-044e-4f63-8af1-51e77d5acce4-etc-machine-id\") pod \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149216 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aca4e1b-044e-4f63-8af1-51e77d5acce4-logs\") pod \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149314 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-scripts\") pod \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149434 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-combined-ca-bundle\") pod \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149549 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data\") pod \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\" (UID: \"5aca4e1b-044e-4f63-8af1-51e77d5acce4\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.149918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aca4e1b-044e-4f63-8af1-51e77d5acce4-logs" (OuterVolumeSpecName: "logs") pod "5aca4e1b-044e-4f63-8af1-51e77d5acce4" (UID: "5aca4e1b-044e-4f63-8af1-51e77d5acce4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.151329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5aca4e1b-044e-4f63-8af1-51e77d5acce4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5aca4e1b-044e-4f63-8af1-51e77d5acce4" (UID: "5aca4e1b-044e-4f63-8af1-51e77d5acce4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.157167 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5aca4e1b-044e-4f63-8af1-51e77d5acce4" (UID: "5aca4e1b-044e-4f63-8af1-51e77d5acce4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.157242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aca4e1b-044e-4f63-8af1-51e77d5acce4-kube-api-access-sscc7" (OuterVolumeSpecName: "kube-api-access-sscc7") pod "5aca4e1b-044e-4f63-8af1-51e77d5acce4" (UID: "5aca4e1b-044e-4f63-8af1-51e77d5acce4"). InnerVolumeSpecName "kube-api-access-sscc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.157834 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-scripts" (OuterVolumeSpecName: "scripts") pod "5aca4e1b-044e-4f63-8af1-51e77d5acce4" (UID: "5aca4e1b-044e-4f63-8af1-51e77d5acce4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.182629 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5aca4e1b-044e-4f63-8af1-51e77d5acce4" (UID: "5aca4e1b-044e-4f63-8af1-51e77d5acce4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.230290 4810 generic.go:334] "Generic (PLEG): container finished" podID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerID="2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55" exitCode=0 Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.230349 4810 generic.go:334] "Generic (PLEG): container finished" podID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerID="c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972" exitCode=143 Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.230372 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.230401 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5aca4e1b-044e-4f63-8af1-51e77d5acce4","Type":"ContainerDied","Data":"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55"} Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.230429 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5aca4e1b-044e-4f63-8af1-51e77d5acce4","Type":"ContainerDied","Data":"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972"} Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.230439 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5aca4e1b-044e-4f63-8af1-51e77d5acce4","Type":"ContainerDied","Data":"e2c3c0f04cb614bafbd025ca6199adad6c4269b0172dd2560fafb504617d8728"} Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.230458 4810 scope.go:117] "RemoveContainer" containerID="2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.233292 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data" (OuterVolumeSpecName: "config-data") pod "5aca4e1b-044e-4f63-8af1-51e77d5acce4" (UID: "5aca4e1b-044e-4f63-8af1-51e77d5acce4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.234889 4810 generic.go:334] "Generic (PLEG): container finished" podID="791f9b40-309e-4266-9962-919281a31ffa" containerID="71222aa26baec746c9bdfce055269e0b6e48395ed1fd372b8dbed9e415120ba7" exitCode=0 Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.236127 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" event={"ID":"791f9b40-309e-4266-9962-919281a31ffa","Type":"ContainerDied","Data":"71222aa26baec746c9bdfce055269e0b6e48395ed1fd372b8dbed9e415120ba7"} Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.238332 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.253714 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.253754 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sscc7\" (UniqueName: \"kubernetes.io/projected/5aca4e1b-044e-4f63-8af1-51e77d5acce4-kube-api-access-sscc7\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.253770 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.253781 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5aca4e1b-044e-4f63-8af1-51e77d5acce4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.253792 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aca4e1b-044e-4f63-8af1-51e77d5acce4-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.253803 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.253812 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aca4e1b-044e-4f63-8af1-51e77d5acce4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.269243 4810 scope.go:117] "RemoveContainer" containerID="c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.309978 4810 scope.go:117] "RemoveContainer" containerID="2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55" Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.312894 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55\": container with ID starting with 2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55 not found: ID does not exist" containerID="2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.312955 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55"} err="failed to get container status \"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55\": rpc error: code = NotFound desc = could not find container \"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55\": container with ID starting with 2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55 not found: ID does not exist" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.313016 4810 scope.go:117] "RemoveContainer" containerID="c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972" Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.319634 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972\": container with ID starting with c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972 not found: ID does not exist" containerID="c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.319694 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972"} err="failed to get container status \"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972\": rpc error: code = NotFound desc = could not find container \"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972\": container with ID starting with c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972 not found: ID does not exist" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.319735 4810 scope.go:117] "RemoveContainer" containerID="2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.320498 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55"} err="failed to get container status \"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55\": rpc error: code = NotFound desc = could not find container \"2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55\": container with ID starting with 2864480fa9706620d673415b376fc3049df038f73fc287be9d0117c85eb2be55 not found: ID does not exist" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.320550 4810 scope.go:117] "RemoveContainer" containerID="c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.320858 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972"} err="failed to get container status \"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972\": rpc error: code = NotFound desc = could not find container \"c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972\": container with ID starting with c9e48da0570ac7f597467f2a066e99ebe40b014fee5279935db9213c6ab95972 not found: ID does not exist" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.356747 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-config\") pod \"791f9b40-309e-4266-9962-919281a31ffa\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.357038 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-swift-storage-0\") pod \"791f9b40-309e-4266-9962-919281a31ffa\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.357088 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-svc\") pod \"791f9b40-309e-4266-9962-919281a31ffa\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.357113 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv4ck\" (UniqueName: \"kubernetes.io/projected/791f9b40-309e-4266-9962-919281a31ffa-kube-api-access-hv4ck\") pod \"791f9b40-309e-4266-9962-919281a31ffa\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.357187 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-nb\") pod \"791f9b40-309e-4266-9962-919281a31ffa\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.357243 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-sb\") pod \"791f9b40-309e-4266-9962-919281a31ffa\" (UID: \"791f9b40-309e-4266-9962-919281a31ffa\") " Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.377386 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791f9b40-309e-4266-9962-919281a31ffa-kube-api-access-hv4ck" (OuterVolumeSpecName: "kube-api-access-hv4ck") pod "791f9b40-309e-4266-9962-919281a31ffa" (UID: "791f9b40-309e-4266-9962-919281a31ffa"). InnerVolumeSpecName "kube-api-access-hv4ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.388367 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.438984 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "791f9b40-309e-4266-9962-919281a31ffa" (UID: "791f9b40-309e-4266-9962-919281a31ffa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.451522 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "791f9b40-309e-4266-9962-919281a31ffa" (UID: "791f9b40-309e-4266-9962-919281a31ffa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.461438 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.461468 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv4ck\" (UniqueName: \"kubernetes.io/projected/791f9b40-309e-4266-9962-919281a31ffa-kube-api-access-hv4ck\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.461481 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.464418 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "791f9b40-309e-4266-9962-919281a31ffa" (UID: "791f9b40-309e-4266-9962-919281a31ffa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.470865 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-67b6b58864-6qdbr"] Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.471114 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-67b6b58864-6qdbr" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-api" containerID="cri-o://f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4" gracePeriod=30 Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.471557 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-67b6b58864-6qdbr" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-httpd" containerID="cri-o://f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf" gracePeriod=30 Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.490531 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-config" (OuterVolumeSpecName: "config") pod "791f9b40-309e-4266-9962-919281a31ffa" (UID: "791f9b40-309e-4266-9962-919281a31ffa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.512781 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "791f9b40-309e-4266-9962-919281a31ffa" (UID: "791f9b40-309e-4266-9962-919281a31ffa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.566308 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.566344 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.566355 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/791f9b40-309e-4266-9962-919281a31ffa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.632451 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.658642 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.748434 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.749591 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.749620 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api" Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.749686 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api-log" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.749695 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api-log" Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.749717 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791f9b40-309e-4266-9962-919281a31ffa" containerName="init" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.749727 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="791f9b40-309e-4266-9962-919281a31ffa" containerName="init" Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.749766 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791f9b40-309e-4266-9962-919281a31ffa" containerName="dnsmasq-dns" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.749774 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="791f9b40-309e-4266-9962-919281a31ffa" containerName="dnsmasq-dns" Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.749796 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api-log" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.749803 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api-log" Oct 08 06:50:00 crc kubenswrapper[4810]: E1008 06:50:00.749838 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.749845 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.750239 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="791f9b40-309e-4266-9962-919281a31ffa" containerName="dnsmasq-dns" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.750346 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.750365 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" containerName="cinder-api-log" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.750388 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.750411 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d38ea1-5f2f-46c0-bdf9-b7f3ed20af34" containerName="barbican-api-log" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.752420 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.756140 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.763066 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.763248 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.772013 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890159 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890290 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890363 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35e33182-d302-417e-8e4e-fec4808047a4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890407 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzvtm\" (UniqueName: \"kubernetes.io/projected/35e33182-d302-417e-8e4e-fec4808047a4-kube-api-access-vzvtm\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890434 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data-custom\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890462 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-scripts\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890514 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e33182-d302-417e-8e4e-fec4808047a4-logs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.890536 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.992511 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993031 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e33182-d302-417e-8e4e-fec4808047a4-logs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993056 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993203 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35e33182-d302-417e-8e4e-fec4808047a4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993240 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzvtm\" (UniqueName: \"kubernetes.io/projected/35e33182-d302-417e-8e4e-fec4808047a4-kube-api-access-vzvtm\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993262 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data-custom\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.993288 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-scripts\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.994203 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e33182-d302-417e-8e4e-fec4808047a4-logs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.994599 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35e33182-d302-417e-8e4e-fec4808047a4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.998735 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:00 crc kubenswrapper[4810]: I1008 06:50:00.999461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.000385 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.000452 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.008733 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-scripts\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.008960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data-custom\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.018738 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzvtm\" (UniqueName: \"kubernetes.io/projected/35e33182-d302-417e-8e4e-fec4808047a4-kube-api-access-vzvtm\") pod \"cinder-api-0\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " pod="openstack/cinder-api-0" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.105890 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.250747 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerStarted","Data":"7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853"} Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.250784 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerStarted","Data":"edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b"} Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.264418 4810 generic.go:334] "Generic (PLEG): container finished" podID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerID="f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf" exitCode=0 Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.264496 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67b6b58864-6qdbr" event={"ID":"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a","Type":"ContainerDied","Data":"f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf"} Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.275892 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.276780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c78787df7-xcmqq" event={"ID":"791f9b40-309e-4266-9962-919281a31ffa","Type":"ContainerDied","Data":"c4f86d7de06ba0010017ba75fa49affc674ca8ee38a5c220c92c55b06940f7ed"} Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.276823 4810 scope.go:117] "RemoveContainer" containerID="71222aa26baec746c9bdfce055269e0b6e48395ed1fd372b8dbed9e415120ba7" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.325508 4810 scope.go:117] "RemoveContainer" containerID="45b68c993c59b693c259ee3bbca5cedcb1d36e22d8d2bb51bfc0b245f89387f8" Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.338256 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c78787df7-xcmqq"] Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.346421 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c78787df7-xcmqq"] Oct 08 06:50:01 crc kubenswrapper[4810]: I1008 06:50:01.672047 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:50:01 crc kubenswrapper[4810]: W1008 06:50:01.673625 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35e33182_d302_417e_8e4e_fec4808047a4.slice/crio-01df55a5848e63b58ce0cc7ad622be8c16ce171c68fd9e1c1842b12ddf8194e4 WatchSource:0}: Error finding container 01df55a5848e63b58ce0cc7ad622be8c16ce171c68fd9e1c1842b12ddf8194e4: Status 404 returned error can't find the container with id 01df55a5848e63b58ce0cc7ad622be8c16ce171c68fd9e1c1842b12ddf8194e4 Oct 08 06:50:02 crc kubenswrapper[4810]: I1008 06:50:02.099134 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aca4e1b-044e-4f63-8af1-51e77d5acce4" path="/var/lib/kubelet/pods/5aca4e1b-044e-4f63-8af1-51e77d5acce4/volumes" Oct 08 06:50:02 crc kubenswrapper[4810]: I1008 06:50:02.101026 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="791f9b40-309e-4266-9962-919281a31ffa" path="/var/lib/kubelet/pods/791f9b40-309e-4266-9962-919281a31ffa/volumes" Oct 08 06:50:02 crc kubenswrapper[4810]: I1008 06:50:02.295981 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"35e33182-d302-417e-8e4e-fec4808047a4","Type":"ContainerStarted","Data":"01df55a5848e63b58ce0cc7ad622be8c16ce171c68fd9e1c1842b12ddf8194e4"} Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.310363 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerStarted","Data":"0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a"} Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.310502 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-central-agent" containerID="cri-o://7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9" gracePeriod=30 Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.311015 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="proxy-httpd" containerID="cri-o://0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a" gracePeriod=30 Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.311118 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-notification-agent" containerID="cri-o://edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b" gracePeriod=30 Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.311143 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.311167 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="sg-core" containerID="cri-o://7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853" gracePeriod=30 Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.318469 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"35e33182-d302-417e-8e4e-fec4808047a4","Type":"ContainerStarted","Data":"8d3bf6f25cf1c45bf48dbb9c1a31f7c04f185f3159a9935da5533577297101f4"} Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.318514 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"35e33182-d302-417e-8e4e-fec4808047a4","Type":"ContainerStarted","Data":"811cbf9580f361194bc54a2a641dd527653e175eccb2b963cfa94e4f09a79794"} Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.318700 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.348016 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.083131677 podStartE2EDuration="12.347985799s" podCreationTimestamp="2025-10-08 06:49:51 +0000 UTC" firstStartedPulling="2025-10-08 06:49:58.23877 +0000 UTC m=+1100.873209750" lastFinishedPulling="2025-10-08 06:50:02.503624132 +0000 UTC m=+1105.138063872" observedRunningTime="2025-10-08 06:50:03.341035077 +0000 UTC m=+1105.975474817" watchObservedRunningTime="2025-10-08 06:50:03.347985799 +0000 UTC m=+1105.982425529" Oct 08 06:50:03 crc kubenswrapper[4810]: I1008 06:50:03.368872 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.368854148 podStartE2EDuration="3.368854148s" podCreationTimestamp="2025-10-08 06:50:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:03.363056227 +0000 UTC m=+1105.997495967" watchObservedRunningTime="2025-10-08 06:50:03.368854148 +0000 UTC m=+1106.003293878" Oct 08 06:50:04 crc kubenswrapper[4810]: I1008 06:50:04.334086 4810 generic.go:334] "Generic (PLEG): container finished" podID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerID="0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a" exitCode=0 Oct 08 06:50:04 crc kubenswrapper[4810]: I1008 06:50:04.334544 4810 generic.go:334] "Generic (PLEG): container finished" podID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerID="7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853" exitCode=2 Oct 08 06:50:04 crc kubenswrapper[4810]: I1008 06:50:04.334558 4810 generic.go:334] "Generic (PLEG): container finished" podID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerID="edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b" exitCode=0 Oct 08 06:50:04 crc kubenswrapper[4810]: I1008 06:50:04.334111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerDied","Data":"0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a"} Oct 08 06:50:04 crc kubenswrapper[4810]: I1008 06:50:04.334679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerDied","Data":"7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853"} Oct 08 06:50:04 crc kubenswrapper[4810]: I1008 06:50:04.334695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerDied","Data":"edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b"} Oct 08 06:50:05 crc kubenswrapper[4810]: I1008 06:50:05.131981 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 06:50:05 crc kubenswrapper[4810]: I1008 06:50:05.194181 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:50:05 crc kubenswrapper[4810]: I1008 06:50:05.341845 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="cinder-scheduler" containerID="cri-o://def11983963f3bac578eb6868e32992fb949f4d82a70c0cdeab56ce2cc363305" gracePeriod=30 Oct 08 06:50:05 crc kubenswrapper[4810]: I1008 06:50:05.342378 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="probe" containerID="cri-o://362be8af4c64a782d1ca3bc7cbfcd409e92cddc7ff4c65e3f4e36cb4913f9d77" gracePeriod=30 Oct 08 06:50:05 crc kubenswrapper[4810]: I1008 06:50:05.851082 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:05.999930 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h5r5\" (UniqueName: \"kubernetes.io/projected/2c7f6b7e-7579-4441-9ec3-6a421478c21c-kube-api-access-2h5r5\") pod \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000139 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-combined-ca-bundle\") pod \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000208 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-run-httpd\") pod \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000300 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-config-data\") pod \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000342 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-log-httpd\") pod \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000382 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-sg-core-conf-yaml\") pod \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-scripts\") pod \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\" (UID: \"2c7f6b7e-7579-4441-9ec3-6a421478c21c\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000698 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2c7f6b7e-7579-4441-9ec3-6a421478c21c" (UID: "2c7f6b7e-7579-4441-9ec3-6a421478c21c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.000953 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2c7f6b7e-7579-4441-9ec3-6a421478c21c" (UID: "2c7f6b7e-7579-4441-9ec3-6a421478c21c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.001826 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.001867 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c7f6b7e-7579-4441-9ec3-6a421478c21c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.008803 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c7f6b7e-7579-4441-9ec3-6a421478c21c-kube-api-access-2h5r5" (OuterVolumeSpecName: "kube-api-access-2h5r5") pod "2c7f6b7e-7579-4441-9ec3-6a421478c21c" (UID: "2c7f6b7e-7579-4441-9ec3-6a421478c21c"). InnerVolumeSpecName "kube-api-access-2h5r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.010749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-scripts" (OuterVolumeSpecName: "scripts") pod "2c7f6b7e-7579-4441-9ec3-6a421478c21c" (UID: "2c7f6b7e-7579-4441-9ec3-6a421478c21c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.038795 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2c7f6b7e-7579-4441-9ec3-6a421478c21c" (UID: "2c7f6b7e-7579-4441-9ec3-6a421478c21c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.104426 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h5r5\" (UniqueName: \"kubernetes.io/projected/2c7f6b7e-7579-4441-9ec3-6a421478c21c-kube-api-access-2h5r5\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.104469 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.104481 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.121379 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c7f6b7e-7579-4441-9ec3-6a421478c21c" (UID: "2c7f6b7e-7579-4441-9ec3-6a421478c21c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.150663 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-config-data" (OuterVolumeSpecName: "config-data") pod "2c7f6b7e-7579-4441-9ec3-6a421478c21c" (UID: "2c7f6b7e-7579-4441-9ec3-6a421478c21c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.206582 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.206627 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7f6b7e-7579-4441-9ec3-6a421478c21c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.350732 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18f0018b_b447_41cb_9fdc_6cd1d8bc38fa.slice/crio-362be8af4c64a782d1ca3bc7cbfcd409e92cddc7ff4c65e3f4e36cb4913f9d77.scope\": RecentStats: unable to find data in memory cache]" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.354342 4810 generic.go:334] "Generic (PLEG): container finished" podID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerID="7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9" exitCode=0 Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.354397 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerDied","Data":"7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9"} Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.354440 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.354459 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c7f6b7e-7579-4441-9ec3-6a421478c21c","Type":"ContainerDied","Data":"9638549ecf6d13608addf6517e520bc46d95ec9ea9d2fa3a5070d815251ebd53"} Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.354484 4810 scope.go:117] "RemoveContainer" containerID="0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.365302 4810 generic.go:334] "Generic (PLEG): container finished" podID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerID="362be8af4c64a782d1ca3bc7cbfcd409e92cddc7ff4c65e3f4e36cb4913f9d77" exitCode=0 Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.365333 4810 generic.go:334] "Generic (PLEG): container finished" podID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerID="def11983963f3bac578eb6868e32992fb949f4d82a70c0cdeab56ce2cc363305" exitCode=0 Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.365360 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa","Type":"ContainerDied","Data":"362be8af4c64a782d1ca3bc7cbfcd409e92cddc7ff4c65e3f4e36cb4913f9d77"} Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.365392 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa","Type":"ContainerDied","Data":"def11983963f3bac578eb6868e32992fb949f4d82a70c0cdeab56ce2cc363305"} Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.410103 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.419690 4810 scope.go:117] "RemoveContainer" containerID="7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.428285 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.469763 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.470259 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="proxy-httpd" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470274 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="proxy-httpd" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.470301 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="sg-core" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470308 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="sg-core" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.470323 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-central-agent" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470329 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-central-agent" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.470345 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-notification-agent" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470351 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-notification-agent" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470538 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-notification-agent" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470561 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="proxy-httpd" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470570 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="sg-core" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.470579 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" containerName="ceilometer-central-agent" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.472309 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.474361 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.475519 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.481881 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.486945 4810 scope.go:117] "RemoveContainer" containerID="edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.550301 4810 scope.go:117] "RemoveContainer" containerID="7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.586442 4810 scope.go:117] "RemoveContainer" containerID="0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.590269 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a\": container with ID starting with 0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a not found: ID does not exist" containerID="0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.590336 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a"} err="failed to get container status \"0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a\": rpc error: code = NotFound desc = could not find container \"0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a\": container with ID starting with 0e4e922b72a405b361f3f82b6b76a32e58c606cdbc07977667cad3e809c6de3a not found: ID does not exist" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.590376 4810 scope.go:117] "RemoveContainer" containerID="7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.594952 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853\": container with ID starting with 7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853 not found: ID does not exist" containerID="7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.594999 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853"} err="failed to get container status \"7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853\": rpc error: code = NotFound desc = could not find container \"7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853\": container with ID starting with 7002a7878b39bfc7ae62352545335b3d269fbda77d8969c2a0f6f87c3f79f853 not found: ID does not exist" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.595026 4810 scope.go:117] "RemoveContainer" containerID="edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.595237 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b\": container with ID starting with edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b not found: ID does not exist" containerID="edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.595259 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b"} err="failed to get container status \"edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b\": rpc error: code = NotFound desc = could not find container \"edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b\": container with ID starting with edc67608f08c0c8903f0cbca2269fb7fbc0e560535a3dbe447b02f8778c0e28b not found: ID does not exist" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.595273 4810 scope.go:117] "RemoveContainer" containerID="7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9" Oct 08 06:50:06 crc kubenswrapper[4810]: E1008 06:50:06.595465 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9\": container with ID starting with 7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9 not found: ID does not exist" containerID="7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.595484 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9"} err="failed to get container status \"7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9\": rpc error: code = NotFound desc = could not find container \"7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9\": container with ID starting with 7bf53ba0475c3e9a41d0d1ff1973f9ca3061eff7ecffc6b62519ca3baa7158d9 not found: ID does not exist" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.616201 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.616267 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-scripts\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.616291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkk9l\" (UniqueName: \"kubernetes.io/projected/13355be7-7143-46b0-a2c8-7886ab7d4f39-kube-api-access-xkk9l\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.616338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-config-data\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.616363 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.616439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-run-httpd\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.616487 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-log-httpd\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.666780 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.719255 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-config-data\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.719316 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.719393 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-run-httpd\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.719439 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-log-httpd\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.719464 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.719493 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-scripts\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.719515 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkk9l\" (UniqueName: \"kubernetes.io/projected/13355be7-7143-46b0-a2c8-7886ab7d4f39-kube-api-access-xkk9l\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.720323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-run-httpd\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.724137 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-log-httpd\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.729337 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.729665 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.729757 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-config-data\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.729821 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-scripts\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.752024 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkk9l\" (UniqueName: \"kubernetes.io/projected/13355be7-7143-46b0-a2c8-7886ab7d4f39-kube-api-access-xkk9l\") pod \"ceilometer-0\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.821023 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data-custom\") pod \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.821151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9mth\" (UniqueName: \"kubernetes.io/projected/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-kube-api-access-l9mth\") pod \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.821175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-combined-ca-bundle\") pod \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.821208 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-etc-machine-id\") pod \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.821267 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-scripts\") pod \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.821302 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data\") pod \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\" (UID: \"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa\") " Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.822220 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" (UID: "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.827701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" (UID: "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.828004 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-kube-api-access-l9mth" (OuterVolumeSpecName: "kube-api-access-l9mth") pod "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" (UID: "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa"). InnerVolumeSpecName "kube-api-access-l9mth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.828346 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-scripts" (OuterVolumeSpecName: "scripts") pod "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" (UID: "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.842913 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.906347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" (UID: "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.927515 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.927555 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9mth\" (UniqueName: \"kubernetes.io/projected/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-kube-api-access-l9mth\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.927567 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.927577 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.927585 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:06 crc kubenswrapper[4810]: I1008 06:50:06.978131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data" (OuterVolumeSpecName: "config-data") pod "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" (UID: "18f0018b-b447-41cb-9fdc-6cd1d8bc38fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.029744 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.071185 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4cvst"] Oct 08 06:50:07 crc kubenswrapper[4810]: E1008 06:50:07.071657 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="cinder-scheduler" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.071671 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="cinder-scheduler" Oct 08 06:50:07 crc kubenswrapper[4810]: E1008 06:50:07.071720 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="probe" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.071727 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="probe" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.071932 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="cinder-scheduler" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.071957 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" containerName="probe" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.072596 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4cvst" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.111152 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4cvst"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.256506 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzpx\" (UniqueName: \"kubernetes.io/projected/315506dc-19f7-4299-b56f-8584a4021a70-kube-api-access-vzzpx\") pod \"nova-api-db-create-4cvst\" (UID: \"315506dc-19f7-4299-b56f-8584a4021a70\") " pod="openstack/nova-api-db-create-4cvst" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.274017 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-4fc5f"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.275286 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fc5f" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.284483 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4fc5f"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.359614 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzpx\" (UniqueName: \"kubernetes.io/projected/315506dc-19f7-4299-b56f-8584a4021a70-kube-api-access-vzzpx\") pod \"nova-api-db-create-4cvst\" (UID: \"315506dc-19f7-4299-b56f-8584a4021a70\") " pod="openstack/nova-api-db-create-4cvst" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.380196 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzpx\" (UniqueName: \"kubernetes.io/projected/315506dc-19f7-4299-b56f-8584a4021a70-kube-api-access-vzzpx\") pod \"nova-api-db-create-4cvst\" (UID: \"315506dc-19f7-4299-b56f-8584a4021a70\") " pod="openstack/nova-api-db-create-4cvst" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.383028 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18f0018b-b447-41cb-9fdc-6cd1d8bc38fa","Type":"ContainerDied","Data":"42e8ce13ce2d8c5e543a4649dc3a29ea43b116d6e1f9378e2482b2b0e9c98bd9"} Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.383077 4810 scope.go:117] "RemoveContainer" containerID="362be8af4c64a782d1ca3bc7cbfcd409e92cddc7ff4c65e3f4e36cb4913f9d77" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.383211 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.410235 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.410849 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4cvst" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.423741 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.432501 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.450327 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:50:07 crc kubenswrapper[4810]: W1008 06:50:07.455382 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13355be7_7143_46b0_a2c8_7886ab7d4f39.slice/crio-4d410ad11a87685bb09a7090501c36aa0c050e14ffbd070074e274c0a414385c WatchSource:0}: Error finding container 4d410ad11a87685bb09a7090501c36aa0c050e14ffbd070074e274c0a414385c: Status 404 returned error can't find the container with id 4d410ad11a87685bb09a7090501c36aa0c050e14ffbd070074e274c0a414385c Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.460319 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.464365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mswt5\" (UniqueName: \"kubernetes.io/projected/4d40d07a-2b98-458f-87d9-de8db4233e85-kube-api-access-mswt5\") pod \"nova-cell0-db-create-4fc5f\" (UID: \"4d40d07a-2b98-458f-87d9-de8db4233e85\") " pod="openstack/nova-cell0-db-create-4fc5f" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.468812 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.493077 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.516253 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jvxdd"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.517607 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jvxdd" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.533443 4810 scope.go:117] "RemoveContainer" containerID="def11983963f3bac578eb6868e32992fb949f4d82a70c0cdeab56ce2cc363305" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.540756 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jvxdd"] Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.568093 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.568174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12678e95-0510-4cd6-86c8-701ea97a247e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.568223 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t692\" (UniqueName: \"kubernetes.io/projected/12678e95-0510-4cd6-86c8-701ea97a247e-kube-api-access-4t692\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.568269 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mswt5\" (UniqueName: \"kubernetes.io/projected/4d40d07a-2b98-458f-87d9-de8db4233e85-kube-api-access-mswt5\") pod \"nova-cell0-db-create-4fc5f\" (UID: \"4d40d07a-2b98-458f-87d9-de8db4233e85\") " pod="openstack/nova-cell0-db-create-4fc5f" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.568319 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.568563 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.568876 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-scripts\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.599453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mswt5\" (UniqueName: \"kubernetes.io/projected/4d40d07a-2b98-458f-87d9-de8db4233e85-kube-api-access-mswt5\") pod \"nova-cell0-db-create-4fc5f\" (UID: \"4d40d07a-2b98-458f-87d9-de8db4233e85\") " pod="openstack/nova-cell0-db-create-4fc5f" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.608931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fc5f" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.671246 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.675513 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.675910 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12678e95-0510-4cd6-86c8-701ea97a247e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.676078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12678e95-0510-4cd6-86c8-701ea97a247e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.676140 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t692\" (UniqueName: \"kubernetes.io/projected/12678e95-0510-4cd6-86c8-701ea97a247e-kube-api-access-4t692\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.676225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.676447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4n24\" (UniqueName: \"kubernetes.io/projected/7be74aea-5896-4b32-8f70-363522e5e7ee-kube-api-access-v4n24\") pod \"nova-cell1-db-create-jvxdd\" (UID: \"7be74aea-5896-4b32-8f70-363522e5e7ee\") " pod="openstack/nova-cell1-db-create-jvxdd" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.676478 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.676694 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-scripts\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.682351 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-scripts\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.696991 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.702845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.722689 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t692\" (UniqueName: \"kubernetes.io/projected/12678e95-0510-4cd6-86c8-701ea97a247e-kube-api-access-4t692\") pod \"cinder-scheduler-0\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " pod="openstack/cinder-scheduler-0" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.783198 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4n24\" (UniqueName: \"kubernetes.io/projected/7be74aea-5896-4b32-8f70-363522e5e7ee-kube-api-access-v4n24\") pod \"nova-cell1-db-create-jvxdd\" (UID: \"7be74aea-5896-4b32-8f70-363522e5e7ee\") " pod="openstack/nova-cell1-db-create-jvxdd" Oct 08 06:50:07 crc kubenswrapper[4810]: I1008 06:50:07.808557 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4n24\" (UniqueName: \"kubernetes.io/projected/7be74aea-5896-4b32-8f70-363522e5e7ee-kube-api-access-v4n24\") pod \"nova-cell1-db-create-jvxdd\" (UID: \"7be74aea-5896-4b32-8f70-363522e5e7ee\") " pod="openstack/nova-cell1-db-create-jvxdd" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.016775 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.021552 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jvxdd" Oct 08 06:50:08 crc kubenswrapper[4810]: W1008 06:50:08.128998 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod315506dc_19f7_4299_b56f_8584a4021a70.slice/crio-7493a4ac6e7e414bd7293537ffe954142bb62ce979faa5ae96dde118cfcd60f4 WatchSource:0}: Error finding container 7493a4ac6e7e414bd7293537ffe954142bb62ce979faa5ae96dde118cfcd60f4: Status 404 returned error can't find the container with id 7493a4ac6e7e414bd7293537ffe954142bb62ce979faa5ae96dde118cfcd60f4 Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.138182 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f0018b-b447-41cb-9fdc-6cd1d8bc38fa" path="/var/lib/kubelet/pods/18f0018b-b447-41cb-9fdc-6cd1d8bc38fa/volumes" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.140213 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c7f6b7e-7579-4441-9ec3-6a421478c21c" path="/var/lib/kubelet/pods/2c7f6b7e-7579-4441-9ec3-6a421478c21c/volumes" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.142828 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4cvst"] Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.166052 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.258372 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4fc5f"] Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.293136 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-httpd-config\") pod \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.293296 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-ovndb-tls-certs\") pod \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.293322 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8psk\" (UniqueName: \"kubernetes.io/projected/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-kube-api-access-q8psk\") pod \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.293466 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-combined-ca-bundle\") pod \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.293505 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-config\") pod \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\" (UID: \"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a\") " Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.300399 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-kube-api-access-q8psk" (OuterVolumeSpecName: "kube-api-access-q8psk") pod "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" (UID: "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a"). InnerVolumeSpecName "kube-api-access-q8psk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.302326 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" (UID: "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.382905 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" (UID: "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.387496 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-config" (OuterVolumeSpecName: "config") pod "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" (UID: "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.395371 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8psk\" (UniqueName: \"kubernetes.io/projected/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-kube-api-access-q8psk\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.395401 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.395414 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.395423 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.409580 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4fc5f" event={"ID":"4d40d07a-2b98-458f-87d9-de8db4233e85","Type":"ContainerStarted","Data":"46eee87a72ddfbe6763ffe853957df11fa22c052eb67d1a5f1cdb69fbd751e44"} Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.422302 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerStarted","Data":"ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f"} Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.422349 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerStarted","Data":"4d410ad11a87685bb09a7090501c36aa0c050e14ffbd070074e274c0a414385c"} Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.425206 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" (UID: "d1edfb10-afed-4da0-aa22-9a1e0ab2c22a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.434389 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4cvst" event={"ID":"315506dc-19f7-4299-b56f-8584a4021a70","Type":"ContainerStarted","Data":"7493a4ac6e7e414bd7293537ffe954142bb62ce979faa5ae96dde118cfcd60f4"} Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.449265 4810 generic.go:334] "Generic (PLEG): container finished" podID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerID="f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4" exitCode=0 Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.449333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67b6b58864-6qdbr" event={"ID":"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a","Type":"ContainerDied","Data":"f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4"} Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.449364 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67b6b58864-6qdbr" event={"ID":"d1edfb10-afed-4da0-aa22-9a1e0ab2c22a","Type":"ContainerDied","Data":"f4825f3141c2fd29848bc9240bb7bb135baf7c08e0e2c4c042838ffcf3a1073c"} Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.449384 4810 scope.go:117] "RemoveContainer" containerID="f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.449531 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67b6b58864-6qdbr" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.497953 4810 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.564612 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-4cvst" podStartSLOduration=1.564591697 podStartE2EDuration="1.564591697s" podCreationTimestamp="2025-10-08 06:50:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:08.456387466 +0000 UTC m=+1111.090827216" watchObservedRunningTime="2025-10-08 06:50:08.564591697 +0000 UTC m=+1111.199031437" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.579048 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.611741 4810 scope.go:117] "RemoveContainer" containerID="f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.708865 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jvxdd"] Oct 08 06:50:08 crc kubenswrapper[4810]: W1008 06:50:08.765478 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7be74aea_5896_4b32_8f70_363522e5e7ee.slice/crio-68e575df69ce6f4f9d9f12b79b55fef8af0844e6c7bb244f1123df1f2f5858fc WatchSource:0}: Error finding container 68e575df69ce6f4f9d9f12b79b55fef8af0844e6c7bb244f1123df1f2f5858fc: Status 404 returned error can't find the container with id 68e575df69ce6f4f9d9f12b79b55fef8af0844e6c7bb244f1123df1f2f5858fc Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.829327 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-67b6b58864-6qdbr"] Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.859804 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-67b6b58864-6qdbr"] Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.864351 4810 scope.go:117] "RemoveContainer" containerID="f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf" Oct 08 06:50:08 crc kubenswrapper[4810]: E1008 06:50:08.865399 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf\": container with ID starting with f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf not found: ID does not exist" containerID="f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.865454 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf"} err="failed to get container status \"f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf\": rpc error: code = NotFound desc = could not find container \"f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf\": container with ID starting with f68f67f3cd66c9e98b837eb82b424cc008b14b54ac0da4c2c5c6e4870db8dadf not found: ID does not exist" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.865501 4810 scope.go:117] "RemoveContainer" containerID="f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4" Oct 08 06:50:08 crc kubenswrapper[4810]: E1008 06:50:08.866731 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4\": container with ID starting with f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4 not found: ID does not exist" containerID="f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4" Oct 08 06:50:08 crc kubenswrapper[4810]: I1008 06:50:08.866809 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4"} err="failed to get container status \"f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4\": rpc error: code = NotFound desc = could not find container \"f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4\": container with ID starting with f7533773b0db213c43d5cdd68fbcc63b9da05e3c03240de1e5768af95cbce5f4 not found: ID does not exist" Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.022445 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.473775 4810 generic.go:334] "Generic (PLEG): container finished" podID="315506dc-19f7-4299-b56f-8584a4021a70" containerID="63b11541d3ac4eab5d5f382a400b96fb2be25d39229d2eacfc7b26bafe86a200" exitCode=0 Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.474320 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4cvst" event={"ID":"315506dc-19f7-4299-b56f-8584a4021a70","Type":"ContainerDied","Data":"63b11541d3ac4eab5d5f382a400b96fb2be25d39229d2eacfc7b26bafe86a200"} Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.523610 4810 generic.go:334] "Generic (PLEG): container finished" podID="4d40d07a-2b98-458f-87d9-de8db4233e85" containerID="1750bec16a054a265008e3c25946acd7c889394cd5239ce5e9246677d0be2833" exitCode=0 Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.524073 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4fc5f" event={"ID":"4d40d07a-2b98-458f-87d9-de8db4233e85","Type":"ContainerDied","Data":"1750bec16a054a265008e3c25946acd7c889394cd5239ce5e9246677d0be2833"} Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.534157 4810 generic.go:334] "Generic (PLEG): container finished" podID="7be74aea-5896-4b32-8f70-363522e5e7ee" containerID="457d3161b5820d3665e9f9681e0892c95a60e78b6401eaba1569cde3e2859c2d" exitCode=0 Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.534272 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jvxdd" event={"ID":"7be74aea-5896-4b32-8f70-363522e5e7ee","Type":"ContainerDied","Data":"457d3161b5820d3665e9f9681e0892c95a60e78b6401eaba1569cde3e2859c2d"} Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.534311 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jvxdd" event={"ID":"7be74aea-5896-4b32-8f70-363522e5e7ee","Type":"ContainerStarted","Data":"68e575df69ce6f4f9d9f12b79b55fef8af0844e6c7bb244f1123df1f2f5858fc"} Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.545475 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerStarted","Data":"b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4"} Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.560800 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"12678e95-0510-4cd6-86c8-701ea97a247e","Type":"ContainerStarted","Data":"f6b4a8db2b3498b1430bdc164af3693743a79260b3c2097cdb4bee62cab40e39"} Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.560862 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"12678e95-0510-4cd6-86c8-701ea97a247e","Type":"ContainerStarted","Data":"b2042103e24421016b0b0fcea6127c5320f9c9997792362bf446b2bf56013c9f"} Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.871343 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.872214 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-log" containerID="cri-o://829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d" gracePeriod=30 Oct 08 06:50:09 crc kubenswrapper[4810]: I1008 06:50:09.872316 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-httpd" containerID="cri-o://986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620" gracePeriod=30 Oct 08 06:50:10 crc kubenswrapper[4810]: I1008 06:50:10.090297 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" path="/var/lib/kubelet/pods/d1edfb10-afed-4da0-aa22-9a1e0ab2c22a/volumes" Oct 08 06:50:10 crc kubenswrapper[4810]: I1008 06:50:10.572468 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"12678e95-0510-4cd6-86c8-701ea97a247e","Type":"ContainerStarted","Data":"81749551b081df35eaa781dfa19fa8b4dcb1258c7b85162b7bf3626674f4ace6"} Oct 08 06:50:10 crc kubenswrapper[4810]: I1008 06:50:10.575580 4810 generic.go:334] "Generic (PLEG): container finished" podID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerID="829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d" exitCode=143 Oct 08 06:50:10 crc kubenswrapper[4810]: I1008 06:50:10.575629 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ad9084b-295c-49da-a76c-aa3cde9ffa29","Type":"ContainerDied","Data":"829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d"} Oct 08 06:50:10 crc kubenswrapper[4810]: I1008 06:50:10.577651 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerStarted","Data":"949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942"} Oct 08 06:50:10 crc kubenswrapper[4810]: I1008 06:50:10.606429 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.606403829 podStartE2EDuration="3.606403829s" podCreationTimestamp="2025-10-08 06:50:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:10.602177082 +0000 UTC m=+1113.236616932" watchObservedRunningTime="2025-10-08 06:50:10.606403829 +0000 UTC m=+1113.240843569" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.057280 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fc5f" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.172076 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4cvst" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.181013 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mswt5\" (UniqueName: \"kubernetes.io/projected/4d40d07a-2b98-458f-87d9-de8db4233e85-kube-api-access-mswt5\") pod \"4d40d07a-2b98-458f-87d9-de8db4233e85\" (UID: \"4d40d07a-2b98-458f-87d9-de8db4233e85\") " Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.190641 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jvxdd" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.286832 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzzpx\" (UniqueName: \"kubernetes.io/projected/315506dc-19f7-4299-b56f-8584a4021a70-kube-api-access-vzzpx\") pod \"315506dc-19f7-4299-b56f-8584a4021a70\" (UID: \"315506dc-19f7-4299-b56f-8584a4021a70\") " Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.286913 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4n24\" (UniqueName: \"kubernetes.io/projected/7be74aea-5896-4b32-8f70-363522e5e7ee-kube-api-access-v4n24\") pod \"7be74aea-5896-4b32-8f70-363522e5e7ee\" (UID: \"7be74aea-5896-4b32-8f70-363522e5e7ee\") " Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.296159 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d40d07a-2b98-458f-87d9-de8db4233e85-kube-api-access-mswt5" (OuterVolumeSpecName: "kube-api-access-mswt5") pod "4d40d07a-2b98-458f-87d9-de8db4233e85" (UID: "4d40d07a-2b98-458f-87d9-de8db4233e85"). InnerVolumeSpecName "kube-api-access-mswt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.296315 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/315506dc-19f7-4299-b56f-8584a4021a70-kube-api-access-vzzpx" (OuterVolumeSpecName: "kube-api-access-vzzpx") pod "315506dc-19f7-4299-b56f-8584a4021a70" (UID: "315506dc-19f7-4299-b56f-8584a4021a70"). InnerVolumeSpecName "kube-api-access-vzzpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.300108 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7be74aea-5896-4b32-8f70-363522e5e7ee-kube-api-access-v4n24" (OuterVolumeSpecName: "kube-api-access-v4n24") pod "7be74aea-5896-4b32-8f70-363522e5e7ee" (UID: "7be74aea-5896-4b32-8f70-363522e5e7ee"). InnerVolumeSpecName "kube-api-access-v4n24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.388915 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mswt5\" (UniqueName: \"kubernetes.io/projected/4d40d07a-2b98-458f-87d9-de8db4233e85-kube-api-access-mswt5\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.388986 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzzpx\" (UniqueName: \"kubernetes.io/projected/315506dc-19f7-4299-b56f-8584a4021a70-kube-api-access-vzzpx\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.389001 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4n24\" (UniqueName: \"kubernetes.io/projected/7be74aea-5896-4b32-8f70-363522e5e7ee-kube-api-access-v4n24\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.602123 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jvxdd" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.602248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jvxdd" event={"ID":"7be74aea-5896-4b32-8f70-363522e5e7ee","Type":"ContainerDied","Data":"68e575df69ce6f4f9d9f12b79b55fef8af0844e6c7bb244f1123df1f2f5858fc"} Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.604688 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68e575df69ce6f4f9d9f12b79b55fef8af0844e6c7bb244f1123df1f2f5858fc" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.624270 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4cvst" event={"ID":"315506dc-19f7-4299-b56f-8584a4021a70","Type":"ContainerDied","Data":"7493a4ac6e7e414bd7293537ffe954142bb62ce979faa5ae96dde118cfcd60f4"} Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.624313 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7493a4ac6e7e414bd7293537ffe954142bb62ce979faa5ae96dde118cfcd60f4" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.624392 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4cvst" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.643738 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fc5f" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.643838 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4fc5f" event={"ID":"4d40d07a-2b98-458f-87d9-de8db4233e85","Type":"ContainerDied","Data":"46eee87a72ddfbe6763ffe853957df11fa22c052eb67d1a5f1cdb69fbd751e44"} Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.643892 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46eee87a72ddfbe6763ffe853957df11fa22c052eb67d1a5f1cdb69fbd751e44" Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.984804 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.985209 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-log" containerID="cri-o://e2fd952082a2646936a99d8f5498c1dc178476cd9cb299a67544c11331778da9" gracePeriod=30 Oct 08 06:50:11 crc kubenswrapper[4810]: I1008 06:50:11.985209 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-httpd" containerID="cri-o://17af0d67e7c101fc6c86a1b89cffcf25c43ff30824d53b0441e8c67f457254be" gracePeriod=30 Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.656036 4810 generic.go:334] "Generic (PLEG): container finished" podID="5632ea60-f549-438f-a244-50f811fefc1a" containerID="e2fd952082a2646936a99d8f5498c1dc178476cd9cb299a67544c11331778da9" exitCode=143 Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.656255 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5632ea60-f549-438f-a244-50f811fefc1a","Type":"ContainerDied","Data":"e2fd952082a2646936a99d8f5498c1dc178476cd9cb299a67544c11331778da9"} Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.660403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerStarted","Data":"f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5"} Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.660563 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-central-agent" containerID="cri-o://ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f" gracePeriod=30 Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.660893 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.661240 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="proxy-httpd" containerID="cri-o://f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5" gracePeriod=30 Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.661293 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="sg-core" containerID="cri-o://949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942" gracePeriod=30 Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.661331 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-notification-agent" containerID="cri-o://b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4" gracePeriod=30 Oct 08 06:50:12 crc kubenswrapper[4810]: I1008 06:50:12.688954 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.641828621 podStartE2EDuration="6.68893627s" podCreationTimestamp="2025-10-08 06:50:06 +0000 UTC" firstStartedPulling="2025-10-08 06:50:07.489820848 +0000 UTC m=+1110.124260588" lastFinishedPulling="2025-10-08 06:50:11.536928497 +0000 UTC m=+1114.171368237" observedRunningTime="2025-10-08 06:50:12.685496355 +0000 UTC m=+1115.319936095" watchObservedRunningTime="2025-10-08 06:50:12.68893627 +0000 UTC m=+1115.323376010" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.017886 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.336721 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.630054 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.689103 4810 generic.go:334] "Generic (PLEG): container finished" podID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerID="986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620" exitCode=0 Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.689170 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ad9084b-295c-49da-a76c-aa3cde9ffa29","Type":"ContainerDied","Data":"986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620"} Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.689197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5ad9084b-295c-49da-a76c-aa3cde9ffa29","Type":"ContainerDied","Data":"b75e1d5216e2d3a26d5eae25932d337b3197c0f0f2868a6f3149cc4b7eee522d"} Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.689215 4810 scope.go:117] "RemoveContainer" containerID="986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.689331 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.712130 4810 generic.go:334] "Generic (PLEG): container finished" podID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerID="f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5" exitCode=0 Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.712174 4810 generic.go:334] "Generic (PLEG): container finished" podID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerID="949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942" exitCode=2 Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.712184 4810 generic.go:334] "Generic (PLEG): container finished" podID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerID="b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4" exitCode=0 Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.712210 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerDied","Data":"f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5"} Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.712238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerDied","Data":"949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942"} Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.712251 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerDied","Data":"b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4"} Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.727138 4810 scope.go:117] "RemoveContainer" containerID="829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.746763 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-combined-ca-bundle\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.746809 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb4x2\" (UniqueName: \"kubernetes.io/projected/5ad9084b-295c-49da-a76c-aa3cde9ffa29-kube-api-access-lb4x2\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.746838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-scripts\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.746867 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-httpd-run\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.746973 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.746990 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.747005 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.747040 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-logs\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.747732 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.751264 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-logs" (OuterVolumeSpecName: "logs") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.775290 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad9084b-295c-49da-a76c-aa3cde9ffa29-kube-api-access-lb4x2" (OuterVolumeSpecName: "kube-api-access-lb4x2") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "kube-api-access-lb4x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.778192 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.779219 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-scripts" (OuterVolumeSpecName: "scripts") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.855654 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.855730 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.855742 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad9084b-295c-49da-a76c-aa3cde9ffa29-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.855751 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb4x2\" (UniqueName: \"kubernetes.io/projected/5ad9084b-295c-49da-a76c-aa3cde9ffa29-kube-api-access-lb4x2\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.855764 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.879378 4810 scope.go:117] "RemoveContainer" containerID="986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620" Oct 08 06:50:13 crc kubenswrapper[4810]: E1008 06:50:13.882260 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620\": container with ID starting with 986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620 not found: ID does not exist" containerID="986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.882395 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620"} err="failed to get container status \"986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620\": rpc error: code = NotFound desc = could not find container \"986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620\": container with ID starting with 986b87cd0b6a336a95c9c5af2fea531a9a96ea70328c5328c8fdb2de10dd7620 not found: ID does not exist" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.882477 4810 scope.go:117] "RemoveContainer" containerID="829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d" Oct 08 06:50:13 crc kubenswrapper[4810]: E1008 06:50:13.891247 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d\": container with ID starting with 829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d not found: ID does not exist" containerID="829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.891308 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d"} err="failed to get container status \"829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d\": rpc error: code = NotFound desc = could not find container \"829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d\": container with ID starting with 829b6d5a046d45619b8a67d6fe7afe491e06682dadd0fe27a28930c2f73ebf0d not found: ID does not exist" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.914421 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.950386 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.956078 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.956131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data" (OuterVolumeSpecName: "config-data") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.957171 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.957295 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data\") pod \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\" (UID: \"5ad9084b-295c-49da-a76c-aa3cde9ffa29\") " Oct 08 06:50:13 crc kubenswrapper[4810]: W1008 06:50:13.957377 4810 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5ad9084b-295c-49da-a76c-aa3cde9ffa29/volumes/kubernetes.io~secret/config-data Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.957531 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data" (OuterVolumeSpecName: "config-data") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: W1008 06:50:13.957747 4810 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5ad9084b-295c-49da-a76c-aa3cde9ffa29/volumes/kubernetes.io~secret/public-tls-certs Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.957840 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5ad9084b-295c-49da-a76c-aa3cde9ffa29" (UID: "5ad9084b-295c-49da-a76c-aa3cde9ffa29"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.958223 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.958324 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.958406 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:13 crc kubenswrapper[4810]: I1008 06:50:13.958480 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad9084b-295c-49da-a76c-aa3cde9ffa29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.044096 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.054123 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.064864 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:50:14 crc kubenswrapper[4810]: E1008 06:50:14.065263 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-log" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.065282 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-log" Oct 08 06:50:14 crc kubenswrapper[4810]: E1008 06:50:14.065293 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-httpd" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.065299 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-httpd" Oct 08 06:50:14 crc kubenswrapper[4810]: E1008 06:50:14.065319 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d40d07a-2b98-458f-87d9-de8db4233e85" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.065325 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d40d07a-2b98-458f-87d9-de8db4233e85" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: E1008 06:50:14.065346 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be74aea-5896-4b32-8f70-363522e5e7ee" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.065352 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be74aea-5896-4b32-8f70-363522e5e7ee" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: E1008 06:50:14.065363 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315506dc-19f7-4299-b56f-8584a4021a70" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.065369 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="315506dc-19f7-4299-b56f-8584a4021a70" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: E1008 06:50:14.065382 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-api" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.065388 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-api" Oct 08 06:50:14 crc kubenswrapper[4810]: E1008 06:50:14.065400 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-httpd" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.065406 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-httpd" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.072338 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-log" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.072395 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" containerName="glance-httpd" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.072415 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be74aea-5896-4b32-8f70-363522e5e7ee" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.072432 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-httpd" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.072443 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1edfb10-afed-4da0-aa22-9a1e0ab2c22a" containerName="neutron-api" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.072469 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="315506dc-19f7-4299-b56f-8584a4021a70" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.072485 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d40d07a-2b98-458f-87d9-de8db4233e85" containerName="mariadb-database-create" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.073892 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.077434 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.077619 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.086257 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad9084b-295c-49da-a76c-aa3cde9ffa29" path="/var/lib/kubelet/pods/5ad9084b-295c-49da-a76c-aa3cde9ffa29/volumes" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.091423 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267128 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khkkw\" (UniqueName: \"kubernetes.io/projected/dced1083-d856-4f53-bb6e-19fc64ba95a5-kube-api-access-khkkw\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267216 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-logs\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267274 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-config-data\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267312 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267378 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267403 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.267680 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-scripts\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.370652 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khkkw\" (UniqueName: \"kubernetes.io/projected/dced1083-d856-4f53-bb6e-19fc64ba95a5-kube-api-access-khkkw\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.371049 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-logs\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.371480 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-logs\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.371641 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-config-data\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.372015 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.372125 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.372197 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.372226 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.372303 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.372352 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-scripts\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.373234 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.376538 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-config-data\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.376759 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.377208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.380449 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-scripts\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.390325 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khkkw\" (UniqueName: \"kubernetes.io/projected/dced1083-d856-4f53-bb6e-19fc64ba95a5-kube-api-access-khkkw\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.412266 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " pod="openstack/glance-default-external-api-0" Oct 08 06:50:14 crc kubenswrapper[4810]: I1008 06:50:14.692605 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.330488 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:50:15 crc kubenswrapper[4810]: W1008 06:50:15.337401 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddced1083_d856_4f53_bb6e_19fc64ba95a5.slice/crio-540bad4204615250877efc93603d9fe908a64a8af2c69739d9838bbce3cab6b5 WatchSource:0}: Error finding container 540bad4204615250877efc93603d9fe908a64a8af2c69739d9838bbce3cab6b5: Status 404 returned error can't find the container with id 540bad4204615250877efc93603d9fe908a64a8af2c69739d9838bbce3cab6b5 Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.737880 4810 generic.go:334] "Generic (PLEG): container finished" podID="5632ea60-f549-438f-a244-50f811fefc1a" containerID="17af0d67e7c101fc6c86a1b89cffcf25c43ff30824d53b0441e8c67f457254be" exitCode=0 Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.738348 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5632ea60-f549-438f-a244-50f811fefc1a","Type":"ContainerDied","Data":"17af0d67e7c101fc6c86a1b89cffcf25c43ff30824d53b0441e8c67f457254be"} Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.738377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5632ea60-f549-438f-a244-50f811fefc1a","Type":"ContainerDied","Data":"e994bf63e6e41b4c7a28ffa2e35ce5fe6d57d7da24822021104f2e58454e1eb6"} Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.738401 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e994bf63e6e41b4c7a28ffa2e35ce5fe6d57d7da24822021104f2e58454e1eb6" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.740014 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dced1083-d856-4f53-bb6e-19fc64ba95a5","Type":"ContainerStarted","Data":"540bad4204615250877efc93603d9fe908a64a8af2c69739d9838bbce3cab6b5"} Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.757254 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.915838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-combined-ca-bundle\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.915942 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-httpd-run\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916025 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-internal-tls-certs\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916124 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kprpx\" (UniqueName: \"kubernetes.io/projected/5632ea60-f549-438f-a244-50f811fefc1a-kube-api-access-kprpx\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916213 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-config-data\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916242 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916262 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-logs\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916349 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-scripts\") pod \"5632ea60-f549-438f-a244-50f811fefc1a\" (UID: \"5632ea60-f549-438f-a244-50f811fefc1a\") " Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916565 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.916788 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.920203 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-logs" (OuterVolumeSpecName: "logs") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.927115 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-scripts" (OuterVolumeSpecName: "scripts") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.929718 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.944139 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5632ea60-f549-438f-a244-50f811fefc1a-kube-api-access-kprpx" (OuterVolumeSpecName: "kube-api-access-kprpx") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "kube-api-access-kprpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.949094 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.976168 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-config-data" (OuterVolumeSpecName: "config-data") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:15 crc kubenswrapper[4810]: I1008 06:50:15.996395 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5632ea60-f549-438f-a244-50f811fefc1a" (UID: "5632ea60-f549-438f-a244-50f811fefc1a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.018730 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.019092 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.019165 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5632ea60-f549-438f-a244-50f811fefc1a-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.019223 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.019286 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.019362 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5632ea60-f549-438f-a244-50f811fefc1a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.019429 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kprpx\" (UniqueName: \"kubernetes.io/projected/5632ea60-f549-438f-a244-50f811fefc1a-kube-api-access-kprpx\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.038398 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.121234 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.752778 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dced1083-d856-4f53-bb6e-19fc64ba95a5","Type":"ContainerStarted","Data":"26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e"} Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.752836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dced1083-d856-4f53-bb6e-19fc64ba95a5","Type":"ContainerStarted","Data":"54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d"} Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.752795 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.775304 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.782325 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.804943 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.804926132 podStartE2EDuration="2.804926132s" podCreationTimestamp="2025-10-08 06:50:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:16.80382204 +0000 UTC m=+1119.438261790" watchObservedRunningTime="2025-10-08 06:50:16.804926132 +0000 UTC m=+1119.439365872" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.820881 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:50:16 crc kubenswrapper[4810]: E1008 06:50:16.821867 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-httpd" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.821894 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-httpd" Oct 08 06:50:16 crc kubenswrapper[4810]: E1008 06:50:16.821909 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-log" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.821920 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-log" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.822169 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-log" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.822196 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5632ea60-f549-438f-a244-50f811fefc1a" containerName="glance-httpd" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.823403 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.825799 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.826808 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.839374 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935127 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935458 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935557 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvp5m\" (UniqueName: \"kubernetes.io/projected/48ca6de6-470e-4226-b67b-e58bf64415e3-kube-api-access-rvp5m\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935601 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935647 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:16 crc kubenswrapper[4810]: I1008 06:50:16.935676 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037146 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037205 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037269 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037305 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvp5m\" (UniqueName: \"kubernetes.io/projected/48ca6de6-470e-4226-b67b-e58bf64415e3-kube-api-access-rvp5m\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037330 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037348 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037365 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.037430 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.038954 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.039076 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.040001 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.046689 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.047648 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.050330 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.056239 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.058317 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvp5m\" (UniqueName: \"kubernetes.io/projected/48ca6de6-470e-4226-b67b-e58bf64415e3-kube-api-access-rvp5m\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.082284 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.143705 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.330538 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-627c-account-create-grrp9"] Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.332105 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-627c-account-create-grrp9" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.340260 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.342270 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-627c-account-create-grrp9"] Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.378324 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.447983 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-log-httpd\") pod \"13355be7-7143-46b0-a2c8-7886ab7d4f39\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.448082 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-config-data\") pod \"13355be7-7143-46b0-a2c8-7886ab7d4f39\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.448187 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-combined-ca-bundle\") pod \"13355be7-7143-46b0-a2c8-7886ab7d4f39\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.448269 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-scripts\") pod \"13355be7-7143-46b0-a2c8-7886ab7d4f39\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.448310 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-sg-core-conf-yaml\") pod \"13355be7-7143-46b0-a2c8-7886ab7d4f39\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.448343 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkk9l\" (UniqueName: \"kubernetes.io/projected/13355be7-7143-46b0-a2c8-7886ab7d4f39-kube-api-access-xkk9l\") pod \"13355be7-7143-46b0-a2c8-7886ab7d4f39\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.448444 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-run-httpd\") pod \"13355be7-7143-46b0-a2c8-7886ab7d4f39\" (UID: \"13355be7-7143-46b0-a2c8-7886ab7d4f39\") " Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.448860 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8zmg\" (UniqueName: \"kubernetes.io/projected/6fe43e84-be58-40c1-8e5e-64547f990174-kube-api-access-h8zmg\") pod \"nova-cell0-627c-account-create-grrp9\" (UID: \"6fe43e84-be58-40c1-8e5e-64547f990174\") " pod="openstack/nova-cell0-627c-account-create-grrp9" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.449799 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "13355be7-7143-46b0-a2c8-7886ab7d4f39" (UID: "13355be7-7143-46b0-a2c8-7886ab7d4f39"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.452869 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "13355be7-7143-46b0-a2c8-7886ab7d4f39" (UID: "13355be7-7143-46b0-a2c8-7886ab7d4f39"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.457930 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13355be7-7143-46b0-a2c8-7886ab7d4f39-kube-api-access-xkk9l" (OuterVolumeSpecName: "kube-api-access-xkk9l") pod "13355be7-7143-46b0-a2c8-7886ab7d4f39" (UID: "13355be7-7143-46b0-a2c8-7886ab7d4f39"). InnerVolumeSpecName "kube-api-access-xkk9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.458095 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-scripts" (OuterVolumeSpecName: "scripts") pod "13355be7-7143-46b0-a2c8-7886ab7d4f39" (UID: "13355be7-7143-46b0-a2c8-7886ab7d4f39"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.481940 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "13355be7-7143-46b0-a2c8-7886ab7d4f39" (UID: "13355be7-7143-46b0-a2c8-7886ab7d4f39"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.518989 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-fa08-account-create-5857z"] Oct 08 06:50:17 crc kubenswrapper[4810]: E1008 06:50:17.519498 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="proxy-httpd" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519510 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="proxy-httpd" Oct 08 06:50:17 crc kubenswrapper[4810]: E1008 06:50:17.519546 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="sg-core" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519552 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="sg-core" Oct 08 06:50:17 crc kubenswrapper[4810]: E1008 06:50:17.519566 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-central-agent" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519572 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-central-agent" Oct 08 06:50:17 crc kubenswrapper[4810]: E1008 06:50:17.519588 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-notification-agent" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519593 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-notification-agent" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519832 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-central-agent" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519845 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="ceilometer-notification-agent" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519863 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="sg-core" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.519885 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerName="proxy-httpd" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.520824 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fa08-account-create-5857z" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.522976 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.530174 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-fa08-account-create-5857z"] Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.551589 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8zmg\" (UniqueName: \"kubernetes.io/projected/6fe43e84-be58-40c1-8e5e-64547f990174-kube-api-access-h8zmg\") pod \"nova-cell0-627c-account-create-grrp9\" (UID: \"6fe43e84-be58-40c1-8e5e-64547f990174\") " pod="openstack/nova-cell0-627c-account-create-grrp9" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.551840 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.551853 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13355be7-7143-46b0-a2c8-7886ab7d4f39-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.551865 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.551876 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.551889 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkk9l\" (UniqueName: \"kubernetes.io/projected/13355be7-7143-46b0-a2c8-7886ab7d4f39-kube-api-access-xkk9l\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.577307 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8zmg\" (UniqueName: \"kubernetes.io/projected/6fe43e84-be58-40c1-8e5e-64547f990174-kube-api-access-h8zmg\") pod \"nova-cell0-627c-account-create-grrp9\" (UID: \"6fe43e84-be58-40c1-8e5e-64547f990174\") " pod="openstack/nova-cell0-627c-account-create-grrp9" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.594115 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13355be7-7143-46b0-a2c8-7886ab7d4f39" (UID: "13355be7-7143-46b0-a2c8-7886ab7d4f39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.606503 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-config-data" (OuterVolumeSpecName: "config-data") pod "13355be7-7143-46b0-a2c8-7886ab7d4f39" (UID: "13355be7-7143-46b0-a2c8-7886ab7d4f39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.654211 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzgpm\" (UniqueName: \"kubernetes.io/projected/ec9c0204-fb65-420a-8917-7fc47453599f-kube-api-access-mzgpm\") pod \"nova-cell1-fa08-account-create-5857z\" (UID: \"ec9c0204-fb65-420a-8917-7fc47453599f\") " pod="openstack/nova-cell1-fa08-account-create-5857z" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.654896 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.654936 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13355be7-7143-46b0-a2c8-7886ab7d4f39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.663048 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-627c-account-create-grrp9" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.757225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzgpm\" (UniqueName: \"kubernetes.io/projected/ec9c0204-fb65-420a-8917-7fc47453599f-kube-api-access-mzgpm\") pod \"nova-cell1-fa08-account-create-5857z\" (UID: \"ec9c0204-fb65-420a-8917-7fc47453599f\") " pod="openstack/nova-cell1-fa08-account-create-5857z" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.768146 4810 generic.go:334] "Generic (PLEG): container finished" podID="13355be7-7143-46b0-a2c8-7886ab7d4f39" containerID="ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f" exitCode=0 Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.768229 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.768297 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerDied","Data":"ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f"} Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.768356 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13355be7-7143-46b0-a2c8-7886ab7d4f39","Type":"ContainerDied","Data":"4d410ad11a87685bb09a7090501c36aa0c050e14ffbd070074e274c0a414385c"} Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.768377 4810 scope.go:117] "RemoveContainer" containerID="f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.790538 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzgpm\" (UniqueName: \"kubernetes.io/projected/ec9c0204-fb65-420a-8917-7fc47453599f-kube-api-access-mzgpm\") pod \"nova-cell1-fa08-account-create-5857z\" (UID: \"ec9c0204-fb65-420a-8917-7fc47453599f\") " pod="openstack/nova-cell1-fa08-account-create-5857z" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.845098 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fa08-account-create-5857z" Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.852169 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:50:17 crc kubenswrapper[4810]: I1008 06:50:17.988510 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.008674 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.010207 4810 scope.go:117] "RemoveContainer" containerID="949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.029220 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.036417 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.047081 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.047263 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.089641 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13355be7-7143-46b0-a2c8-7886ab7d4f39" path="/var/lib/kubelet/pods/13355be7-7143-46b0-a2c8-7886ab7d4f39/volumes" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.090804 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5632ea60-f549-438f-a244-50f811fefc1a" path="/var/lib/kubelet/pods/5632ea60-f549-438f-a244-50f811fefc1a/volumes" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.091644 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.128953 4810 scope.go:117] "RemoveContainer" containerID="b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.170548 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.170852 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-run-httpd\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.171006 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-config-data\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.171116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9nmb\" (UniqueName: \"kubernetes.io/projected/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-kube-api-access-q9nmb\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.171607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.172061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-scripts\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.172282 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-log-httpd\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.181041 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-fa08-account-create-5857z"] Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.197239 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-627c-account-create-grrp9"] Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.202186 4810 scope.go:117] "RemoveContainer" containerID="ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f" Oct 08 06:50:18 crc kubenswrapper[4810]: W1008 06:50:18.228122 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fe43e84_be58_40c1_8e5e_64547f990174.slice/crio-9c5fdfa4192c00c18b3a472bbc5ce6dd5da628344c7f327b8f3d9d9b8e511741 WatchSource:0}: Error finding container 9c5fdfa4192c00c18b3a472bbc5ce6dd5da628344c7f327b8f3d9d9b8e511741: Status 404 returned error can't find the container with id 9c5fdfa4192c00c18b3a472bbc5ce6dd5da628344c7f327b8f3d9d9b8e511741 Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.239985 4810 scope.go:117] "RemoveContainer" containerID="f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5" Oct 08 06:50:18 crc kubenswrapper[4810]: E1008 06:50:18.241937 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5\": container with ID starting with f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5 not found: ID does not exist" containerID="f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.242015 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5"} err="failed to get container status \"f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5\": rpc error: code = NotFound desc = could not find container \"f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5\": container with ID starting with f8385080929b4fdc425df0c4540c56dd7adeac4ad60b313e50ab57a161b357d5 not found: ID does not exist" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.242047 4810 scope.go:117] "RemoveContainer" containerID="949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942" Oct 08 06:50:18 crc kubenswrapper[4810]: E1008 06:50:18.247457 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942\": container with ID starting with 949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942 not found: ID does not exist" containerID="949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.247519 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942"} err="failed to get container status \"949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942\": rpc error: code = NotFound desc = could not find container \"949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942\": container with ID starting with 949e0194804dc35f9619697001b6d0d1805c0035b98aa8e95729c3cc6acfd942 not found: ID does not exist" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.247560 4810 scope.go:117] "RemoveContainer" containerID="b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4" Oct 08 06:50:18 crc kubenswrapper[4810]: E1008 06:50:18.248318 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4\": container with ID starting with b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4 not found: ID does not exist" containerID="b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.248444 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4"} err="failed to get container status \"b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4\": rpc error: code = NotFound desc = could not find container \"b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4\": container with ID starting with b04ade1df5c67d18b2db702fc01f6c6ab6bdf949d63e72504009b2f9f9dd72b4 not found: ID does not exist" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.248552 4810 scope.go:117] "RemoveContainer" containerID="ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f" Oct 08 06:50:18 crc kubenswrapper[4810]: E1008 06:50:18.248990 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f\": container with ID starting with ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f not found: ID does not exist" containerID="ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.249036 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f"} err="failed to get container status \"ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f\": rpc error: code = NotFound desc = could not find container \"ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f\": container with ID starting with ec04f72f50a7af2d66c826958c486d26abe4dd39b16478a8fbee47769af5ab8f not found: ID does not exist" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.277197 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-config-data\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.277295 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9nmb\" (UniqueName: \"kubernetes.io/projected/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-kube-api-access-q9nmb\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.277364 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.277433 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-scripts\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.277467 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-log-httpd\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.277515 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.277581 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-run-httpd\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.278188 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-run-httpd\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.279687 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-log-httpd\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.287551 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-scripts\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.291200 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.293536 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.294401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-config-data\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.309074 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9nmb\" (UniqueName: \"kubernetes.io/projected/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-kube-api-access-q9nmb\") pod \"ceilometer-0\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.374176 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.391169 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.767020 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.850375 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec9c0204-fb65-420a-8917-7fc47453599f" containerID="949b16b370e9288b713511d19a1e607e09e78c34f767d6c24c5b4e75b772a646" exitCode=0 Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.850492 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fa08-account-create-5857z" event={"ID":"ec9c0204-fb65-420a-8917-7fc47453599f","Type":"ContainerDied","Data":"949b16b370e9288b713511d19a1e607e09e78c34f767d6c24c5b4e75b772a646"} Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.850544 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fa08-account-create-5857z" event={"ID":"ec9c0204-fb65-420a-8917-7fc47453599f","Type":"ContainerStarted","Data":"e0aaf08d3718a17ad577afb14f75cb62def5c2fb67f13169d1ca942fc1fc16ce"} Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.860304 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48ca6de6-470e-4226-b67b-e58bf64415e3","Type":"ContainerStarted","Data":"b0c365df855d6b728d0caeb86ad0f03ae13282564f65afa5a26564e04b85aa9a"} Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.869821 4810 generic.go:334] "Generic (PLEG): container finished" podID="6fe43e84-be58-40c1-8e5e-64547f990174" containerID="af751edb5dd316cd9dd468c60b6235482bce1444060e49de080862ecf361ab3f" exitCode=0 Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.869874 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-627c-account-create-grrp9" event={"ID":"6fe43e84-be58-40c1-8e5e-64547f990174","Type":"ContainerDied","Data":"af751edb5dd316cd9dd468c60b6235482bce1444060e49de080862ecf361ab3f"} Oct 08 06:50:18 crc kubenswrapper[4810]: I1008 06:50:18.869901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-627c-account-create-grrp9" event={"ID":"6fe43e84-be58-40c1-8e5e-64547f990174","Type":"ContainerStarted","Data":"9c5fdfa4192c00c18b3a472bbc5ce6dd5da628344c7f327b8f3d9d9b8e511741"} Oct 08 06:50:19 crc kubenswrapper[4810]: I1008 06:50:19.431907 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:50:19 crc kubenswrapper[4810]: I1008 06:50:19.432397 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:50:19 crc kubenswrapper[4810]: I1008 06:50:19.884951 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48ca6de6-470e-4226-b67b-e58bf64415e3","Type":"ContainerStarted","Data":"ed04a8db57a733ab9675e113837de747b29cbd4384df9603ebc90167bf592eca"} Oct 08 06:50:19 crc kubenswrapper[4810]: I1008 06:50:19.885439 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48ca6de6-470e-4226-b67b-e58bf64415e3","Type":"ContainerStarted","Data":"47e85962732a13a5f5460d0367ec3af0533b5a5617841d2c9fda2a5452a2fbdb"} Oct 08 06:50:19 crc kubenswrapper[4810]: I1008 06:50:19.889489 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerStarted","Data":"6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649"} Oct 08 06:50:19 crc kubenswrapper[4810]: I1008 06:50:19.889550 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerStarted","Data":"e3e6e4943821929561bfd394b673e3a2b76852b12776ecb26321cc073bf2248e"} Oct 08 06:50:19 crc kubenswrapper[4810]: I1008 06:50:19.928601 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.9285769200000002 podStartE2EDuration="3.92857692s" podCreationTimestamp="2025-10-08 06:50:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:19.912880534 +0000 UTC m=+1122.547320274" watchObservedRunningTime="2025-10-08 06:50:19.92857692 +0000 UTC m=+1122.563016660" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.427198 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-627c-account-create-grrp9" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.436999 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fa08-account-create-5857z" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.556690 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8zmg\" (UniqueName: \"kubernetes.io/projected/6fe43e84-be58-40c1-8e5e-64547f990174-kube-api-access-h8zmg\") pod \"6fe43e84-be58-40c1-8e5e-64547f990174\" (UID: \"6fe43e84-be58-40c1-8e5e-64547f990174\") " Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.556839 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzgpm\" (UniqueName: \"kubernetes.io/projected/ec9c0204-fb65-420a-8917-7fc47453599f-kube-api-access-mzgpm\") pod \"ec9c0204-fb65-420a-8917-7fc47453599f\" (UID: \"ec9c0204-fb65-420a-8917-7fc47453599f\") " Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.563013 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec9c0204-fb65-420a-8917-7fc47453599f-kube-api-access-mzgpm" (OuterVolumeSpecName: "kube-api-access-mzgpm") pod "ec9c0204-fb65-420a-8917-7fc47453599f" (UID: "ec9c0204-fb65-420a-8917-7fc47453599f"). InnerVolumeSpecName "kube-api-access-mzgpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.563123 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fe43e84-be58-40c1-8e5e-64547f990174-kube-api-access-h8zmg" (OuterVolumeSpecName: "kube-api-access-h8zmg") pod "6fe43e84-be58-40c1-8e5e-64547f990174" (UID: "6fe43e84-be58-40c1-8e5e-64547f990174"). InnerVolumeSpecName "kube-api-access-h8zmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.658745 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzgpm\" (UniqueName: \"kubernetes.io/projected/ec9c0204-fb65-420a-8917-7fc47453599f-kube-api-access-mzgpm\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.658780 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8zmg\" (UniqueName: \"kubernetes.io/projected/6fe43e84-be58-40c1-8e5e-64547f990174-kube-api-access-h8zmg\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.900954 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerStarted","Data":"f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae"} Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.902510 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-627c-account-create-grrp9" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.902662 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-627c-account-create-grrp9" event={"ID":"6fe43e84-be58-40c1-8e5e-64547f990174","Type":"ContainerDied","Data":"9c5fdfa4192c00c18b3a472bbc5ce6dd5da628344c7f327b8f3d9d9b8e511741"} Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.902701 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c5fdfa4192c00c18b3a472bbc5ce6dd5da628344c7f327b8f3d9d9b8e511741" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.908578 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fa08-account-create-5857z" Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.908578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fa08-account-create-5857z" event={"ID":"ec9c0204-fb65-420a-8917-7fc47453599f","Type":"ContainerDied","Data":"e0aaf08d3718a17ad577afb14f75cb62def5c2fb67f13169d1ca942fc1fc16ce"} Oct 08 06:50:20 crc kubenswrapper[4810]: I1008 06:50:20.908641 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0aaf08d3718a17ad577afb14f75cb62def5c2fb67f13169d1ca942fc1fc16ce" Oct 08 06:50:21 crc kubenswrapper[4810]: I1008 06:50:21.920693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerStarted","Data":"dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4"} Oct 08 06:50:22 crc kubenswrapper[4810]: I1008 06:50:22.998929 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-77482"] Oct 08 06:50:23 crc kubenswrapper[4810]: E1008 06:50:22.999829 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec9c0204-fb65-420a-8917-7fc47453599f" containerName="mariadb-account-create" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:22.999849 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec9c0204-fb65-420a-8917-7fc47453599f" containerName="mariadb-account-create" Oct 08 06:50:23 crc kubenswrapper[4810]: E1008 06:50:22.999871 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe43e84-be58-40c1-8e5e-64547f990174" containerName="mariadb-account-create" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:22.999879 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe43e84-be58-40c1-8e5e-64547f990174" containerName="mariadb-account-create" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.000172 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe43e84-be58-40c1-8e5e-64547f990174" containerName="mariadb-account-create" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.000200 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec9c0204-fb65-420a-8917-7fc47453599f" containerName="mariadb-account-create" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.001389 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.003140 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vtvjz" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.004109 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.004432 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.026563 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-77482"] Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.110904 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.110995 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lwd2\" (UniqueName: \"kubernetes.io/projected/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-kube-api-access-4lwd2\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.111687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-config-data\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.111871 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-scripts\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.214403 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-config-data\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.214489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-scripts\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.214519 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.214548 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lwd2\" (UniqueName: \"kubernetes.io/projected/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-kube-api-access-4lwd2\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.223359 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-scripts\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.223446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.240234 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-config-data\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.242410 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lwd2\" (UniqueName: \"kubernetes.io/projected/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-kube-api-access-4lwd2\") pod \"nova-cell0-conductor-db-sync-77482\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.329219 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.944443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerStarted","Data":"1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51"} Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.945149 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 06:50:23 crc kubenswrapper[4810]: I1008 06:50:23.980016 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.549010349 podStartE2EDuration="6.979991139s" podCreationTimestamp="2025-10-08 06:50:17 +0000 UTC" firstStartedPulling="2025-10-08 06:50:18.829891736 +0000 UTC m=+1121.464331486" lastFinishedPulling="2025-10-08 06:50:23.260872536 +0000 UTC m=+1125.895312276" observedRunningTime="2025-10-08 06:50:23.972059329 +0000 UTC m=+1126.606499059" watchObservedRunningTime="2025-10-08 06:50:23.979991139 +0000 UTC m=+1126.614430879" Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.014131 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-77482"] Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.693123 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.695393 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.738526 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.770495 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.954437 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-77482" event={"ID":"f7375a67-1872-4a18-a8b6-04cc8a6f06b8","Type":"ContainerStarted","Data":"ad52c257e232397edfe6378df53970737924ec4d54b8d71a32048766b50ad6b2"} Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.955028 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 06:50:24 crc kubenswrapper[4810]: I1008 06:50:24.955055 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 06:50:26 crc kubenswrapper[4810]: I1008 06:50:26.973716 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 06:50:26 crc kubenswrapper[4810]: I1008 06:50:26.974634 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.052984 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.144905 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.145059 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.201989 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.216407 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.247663 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-29ac-account-create-ngrh8"] Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.250245 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-29ac-account-create-ngrh8" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.254654 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.262687 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-29ac-account-create-ngrh8"] Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.306744 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dct5z\" (UniqueName: \"kubernetes.io/projected/01ee716b-7145-415d-a330-d5f9f5c63f36-kube-api-access-dct5z\") pod \"nova-api-29ac-account-create-ngrh8\" (UID: \"01ee716b-7145-415d-a330-d5f9f5c63f36\") " pod="openstack/nova-api-29ac-account-create-ngrh8" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.410848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dct5z\" (UniqueName: \"kubernetes.io/projected/01ee716b-7145-415d-a330-d5f9f5c63f36-kube-api-access-dct5z\") pod \"nova-api-29ac-account-create-ngrh8\" (UID: \"01ee716b-7145-415d-a330-d5f9f5c63f36\") " pod="openstack/nova-api-29ac-account-create-ngrh8" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.445208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dct5z\" (UniqueName: \"kubernetes.io/projected/01ee716b-7145-415d-a330-d5f9f5c63f36-kube-api-access-dct5z\") pod \"nova-api-29ac-account-create-ngrh8\" (UID: \"01ee716b-7145-415d-a330-d5f9f5c63f36\") " pod="openstack/nova-api-29ac-account-create-ngrh8" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.605411 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-29ac-account-create-ngrh8" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.984400 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:27 crc kubenswrapper[4810]: I1008 06:50:27.984456 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:30 crc kubenswrapper[4810]: I1008 06:50:30.288040 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:30 crc kubenswrapper[4810]: I1008 06:50:30.288659 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 06:50:30 crc kubenswrapper[4810]: I1008 06:50:30.319436 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 06:50:31 crc kubenswrapper[4810]: I1008 06:50:31.896952 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-29ac-account-create-ngrh8"] Oct 08 06:50:31 crc kubenswrapper[4810]: W1008 06:50:31.897574 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ee716b_7145_415d_a330_d5f9f5c63f36.slice/crio-e67afd5d403ac2aaa1a719a44e4215e7329060e6cd06cd720bfa354f43d50d50 WatchSource:0}: Error finding container e67afd5d403ac2aaa1a719a44e4215e7329060e6cd06cd720bfa354f43d50d50: Status 404 returned error can't find the container with id e67afd5d403ac2aaa1a719a44e4215e7329060e6cd06cd720bfa354f43d50d50 Oct 08 06:50:32 crc kubenswrapper[4810]: I1008 06:50:32.047778 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-77482" event={"ID":"f7375a67-1872-4a18-a8b6-04cc8a6f06b8","Type":"ContainerStarted","Data":"9241825651b89e0de120cb13a697f373ba57f94026cdad88f2152713fec8f376"} Oct 08 06:50:32 crc kubenswrapper[4810]: I1008 06:50:32.050459 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-29ac-account-create-ngrh8" event={"ID":"01ee716b-7145-415d-a330-d5f9f5c63f36","Type":"ContainerStarted","Data":"e67afd5d403ac2aaa1a719a44e4215e7329060e6cd06cd720bfa354f43d50d50"} Oct 08 06:50:32 crc kubenswrapper[4810]: I1008 06:50:32.076086 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-77482" podStartSLOduration=2.593187622 podStartE2EDuration="10.076054059s" podCreationTimestamp="2025-10-08 06:50:22 +0000 UTC" firstStartedPulling="2025-10-08 06:50:24.018398425 +0000 UTC m=+1126.652838165" lastFinishedPulling="2025-10-08 06:50:31.501264862 +0000 UTC m=+1134.135704602" observedRunningTime="2025-10-08 06:50:32.068302974 +0000 UTC m=+1134.702742734" watchObservedRunningTime="2025-10-08 06:50:32.076054059 +0000 UTC m=+1134.710493819" Oct 08 06:50:33 crc kubenswrapper[4810]: I1008 06:50:33.062721 4810 generic.go:334] "Generic (PLEG): container finished" podID="01ee716b-7145-415d-a330-d5f9f5c63f36" containerID="811375e4fddcc6b30fb161f68a76625844e575fef371e71934b1514a60a8c6a3" exitCode=0 Oct 08 06:50:33 crc kubenswrapper[4810]: I1008 06:50:33.065820 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-29ac-account-create-ngrh8" event={"ID":"01ee716b-7145-415d-a330-d5f9f5c63f36","Type":"ContainerDied","Data":"811375e4fddcc6b30fb161f68a76625844e575fef371e71934b1514a60a8c6a3"} Oct 08 06:50:34 crc kubenswrapper[4810]: I1008 06:50:34.432464 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-29ac-account-create-ngrh8" Oct 08 06:50:34 crc kubenswrapper[4810]: I1008 06:50:34.588020 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dct5z\" (UniqueName: \"kubernetes.io/projected/01ee716b-7145-415d-a330-d5f9f5c63f36-kube-api-access-dct5z\") pod \"01ee716b-7145-415d-a330-d5f9f5c63f36\" (UID: \"01ee716b-7145-415d-a330-d5f9f5c63f36\") " Oct 08 06:50:34 crc kubenswrapper[4810]: I1008 06:50:34.598420 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ee716b-7145-415d-a330-d5f9f5c63f36-kube-api-access-dct5z" (OuterVolumeSpecName: "kube-api-access-dct5z") pod "01ee716b-7145-415d-a330-d5f9f5c63f36" (UID: "01ee716b-7145-415d-a330-d5f9f5c63f36"). InnerVolumeSpecName "kube-api-access-dct5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:34 crc kubenswrapper[4810]: I1008 06:50:34.690770 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dct5z\" (UniqueName: \"kubernetes.io/projected/01ee716b-7145-415d-a330-d5f9f5c63f36-kube-api-access-dct5z\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:35 crc kubenswrapper[4810]: I1008 06:50:35.087119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-29ac-account-create-ngrh8" event={"ID":"01ee716b-7145-415d-a330-d5f9f5c63f36","Type":"ContainerDied","Data":"e67afd5d403ac2aaa1a719a44e4215e7329060e6cd06cd720bfa354f43d50d50"} Oct 08 06:50:35 crc kubenswrapper[4810]: I1008 06:50:35.087539 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e67afd5d403ac2aaa1a719a44e4215e7329060e6cd06cd720bfa354f43d50d50" Oct 08 06:50:35 crc kubenswrapper[4810]: I1008 06:50:35.087198 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-29ac-account-create-ngrh8" Oct 08 06:50:36 crc kubenswrapper[4810]: I1008 06:50:36.957091 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:36 crc kubenswrapper[4810]: I1008 06:50:36.957608 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-central-agent" containerID="cri-o://6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649" gracePeriod=30 Oct 08 06:50:36 crc kubenswrapper[4810]: I1008 06:50:36.957687 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="sg-core" containerID="cri-o://dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4" gracePeriod=30 Oct 08 06:50:36 crc kubenswrapper[4810]: I1008 06:50:36.957720 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-notification-agent" containerID="cri-o://f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae" gracePeriod=30 Oct 08 06:50:36 crc kubenswrapper[4810]: I1008 06:50:36.957886 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="proxy-httpd" containerID="cri-o://1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51" gracePeriod=30 Oct 08 06:50:36 crc kubenswrapper[4810]: I1008 06:50:36.966489 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 06:50:37 crc kubenswrapper[4810]: I1008 06:50:37.112589 4810 generic.go:334] "Generic (PLEG): container finished" podID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerID="1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51" exitCode=0 Oct 08 06:50:37 crc kubenswrapper[4810]: I1008 06:50:37.112634 4810 generic.go:334] "Generic (PLEG): container finished" podID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerID="dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4" exitCode=2 Oct 08 06:50:37 crc kubenswrapper[4810]: I1008 06:50:37.112659 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerDied","Data":"1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51"} Oct 08 06:50:37 crc kubenswrapper[4810]: I1008 06:50:37.112690 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerDied","Data":"dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4"} Oct 08 06:50:37 crc kubenswrapper[4810]: E1008 06:50:37.166432 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40caa5ae_1fdd_4103_b2d5_c9d0ed63e832.slice/crio-1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40caa5ae_1fdd_4103_b2d5_c9d0ed63e832.slice/crio-dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40caa5ae_1fdd_4103_b2d5_c9d0ed63e832.slice/crio-conmon-dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4.scope\": RecentStats: unable to find data in memory cache]" Oct 08 06:50:38 crc kubenswrapper[4810]: I1008 06:50:38.125977 4810 generic.go:334] "Generic (PLEG): container finished" podID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerID="6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649" exitCode=0 Oct 08 06:50:38 crc kubenswrapper[4810]: I1008 06:50:38.126492 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerDied","Data":"6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649"} Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.131028 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.139747 4810 generic.go:334] "Generic (PLEG): container finished" podID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerID="f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae" exitCode=0 Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.139812 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerDied","Data":"f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae"} Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.139842 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.139855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832","Type":"ContainerDied","Data":"e3e6e4943821929561bfd394b673e3a2b76852b12776ecb26321cc073bf2248e"} Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.139882 4810 scope.go:117] "RemoveContainer" containerID="1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.162063 4810 scope.go:117] "RemoveContainer" containerID="dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.188108 4810 scope.go:117] "RemoveContainer" containerID="f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.210077 4810 scope.go:117] "RemoveContainer" containerID="6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.231016 4810 scope.go:117] "RemoveContainer" containerID="1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.231530 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51\": container with ID starting with 1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51 not found: ID does not exist" containerID="1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.231577 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51"} err="failed to get container status \"1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51\": rpc error: code = NotFound desc = could not find container \"1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51\": container with ID starting with 1e103a2dc8936d45aaf77603ee401f792c73f22d499070e43180e67ffcc9ce51 not found: ID does not exist" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.231607 4810 scope.go:117] "RemoveContainer" containerID="dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.232216 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4\": container with ID starting with dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4 not found: ID does not exist" containerID="dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.232238 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4"} err="failed to get container status \"dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4\": rpc error: code = NotFound desc = could not find container \"dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4\": container with ID starting with dc4028c472a439e4db00a109422becbce86677a1dfe3e7f7e4a94f81aa40c3b4 not found: ID does not exist" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.232253 4810 scope.go:117] "RemoveContainer" containerID="f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.232603 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae\": container with ID starting with f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae not found: ID does not exist" containerID="f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.232627 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae"} err="failed to get container status \"f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae\": rpc error: code = NotFound desc = could not find container \"f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae\": container with ID starting with f35a0cc2524b33d6752df53e335560dd97c7eba8ae07199b30cc9abf490346ae not found: ID does not exist" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.232641 4810 scope.go:117] "RemoveContainer" containerID="6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.232891 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649\": container with ID starting with 6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649 not found: ID does not exist" containerID="6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.232915 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649"} err="failed to get container status \"6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649\": rpc error: code = NotFound desc = could not find container \"6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649\": container with ID starting with 6dc98dce146e8f890083f189dcb09fe6ad27ce77b802f57b9341e1859a63e649 not found: ID does not exist" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.280590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9nmb\" (UniqueName: \"kubernetes.io/projected/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-kube-api-access-q9nmb\") pod \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.280641 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-scripts\") pod \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.280768 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-combined-ca-bundle\") pod \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.280836 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-run-httpd\") pod \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.280884 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-log-httpd\") pod \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.280929 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-sg-core-conf-yaml\") pod \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.281059 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-config-data\") pod \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\" (UID: \"40caa5ae-1fdd-4103-b2d5-c9d0ed63e832\") " Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.281427 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" (UID: "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.281443 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" (UID: "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.282071 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.282097 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.286783 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-kube-api-access-q9nmb" (OuterVolumeSpecName: "kube-api-access-q9nmb") pod "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" (UID: "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832"). InnerVolumeSpecName "kube-api-access-q9nmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.301863 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-scripts" (OuterVolumeSpecName: "scripts") pod "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" (UID: "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.309528 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" (UID: "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.370209 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" (UID: "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.384530 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9nmb\" (UniqueName: \"kubernetes.io/projected/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-kube-api-access-q9nmb\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.384588 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.384607 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.384625 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.418064 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-config-data" (OuterVolumeSpecName: "config-data") pod "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" (UID: "40caa5ae-1fdd-4103-b2d5-c9d0ed63e832"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.478886 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.487396 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.491064 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.506937 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.507321 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ee716b-7145-415d-a330-d5f9f5c63f36" containerName="mariadb-account-create" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507335 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ee716b-7145-415d-a330-d5f9f5c63f36" containerName="mariadb-account-create" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.507353 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="proxy-httpd" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507359 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="proxy-httpd" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.507374 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="sg-core" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507381 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="sg-core" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.507402 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-notification-agent" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507408 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-notification-agent" Oct 08 06:50:39 crc kubenswrapper[4810]: E1008 06:50:39.507421 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-central-agent" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507428 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-central-agent" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507607 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ee716b-7145-415d-a330-d5f9f5c63f36" containerName="mariadb-account-create" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507635 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="sg-core" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507644 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-central-agent" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507651 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="ceilometer-notification-agent" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.507664 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" containerName="proxy-httpd" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.509335 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.511625 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.512248 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.572614 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.690997 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-log-httpd\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.691408 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-config-data\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.691536 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.691664 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-run-httpd\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.691757 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.691839 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4mks\" (UniqueName: \"kubernetes.io/projected/e06a2355-ff17-4cb4-bd37-63387901529e-kube-api-access-v4mks\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.691947 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-scripts\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.793901 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-log-httpd\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794030 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-config-data\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794071 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-run-httpd\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794187 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4mks\" (UniqueName: \"kubernetes.io/projected/e06a2355-ff17-4cb4-bd37-63387901529e-kube-api-access-v4mks\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-scripts\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794389 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-log-httpd\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.794592 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-run-httpd\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.800506 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.800792 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.801316 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-config-data\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.801366 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-scripts\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.813750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4mks\" (UniqueName: \"kubernetes.io/projected/e06a2355-ff17-4cb4-bd37-63387901529e-kube-api-access-v4mks\") pod \"ceilometer-0\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " pod="openstack/ceilometer-0" Oct 08 06:50:39 crc kubenswrapper[4810]: I1008 06:50:39.889484 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:40 crc kubenswrapper[4810]: I1008 06:50:40.101936 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40caa5ae-1fdd-4103-b2d5-c9d0ed63e832" path="/var/lib/kubelet/pods/40caa5ae-1fdd-4103-b2d5-c9d0ed63e832/volumes" Oct 08 06:50:40 crc kubenswrapper[4810]: I1008 06:50:40.383213 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:41 crc kubenswrapper[4810]: I1008 06:50:41.185700 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerStarted","Data":"b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47"} Oct 08 06:50:41 crc kubenswrapper[4810]: I1008 06:50:41.186175 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerStarted","Data":"cff6de5e294ab4401219b8934a4ba3a4f2024b615010017771ec09573b1fcabd"} Oct 08 06:50:41 crc kubenswrapper[4810]: I1008 06:50:41.376154 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:50:41 crc kubenswrapper[4810]: I1008 06:50:41.376815 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="372bbd1e-a3e6-4f87-b824-4ee000006789" containerName="kube-state-metrics" containerID="cri-o://d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de" gracePeriod=30 Oct 08 06:50:41 crc kubenswrapper[4810]: I1008 06:50:41.819734 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:50:41 crc kubenswrapper[4810]: I1008 06:50:41.952949 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq8jv\" (UniqueName: \"kubernetes.io/projected/372bbd1e-a3e6-4f87-b824-4ee000006789-kube-api-access-sq8jv\") pod \"372bbd1e-a3e6-4f87-b824-4ee000006789\" (UID: \"372bbd1e-a3e6-4f87-b824-4ee000006789\") " Oct 08 06:50:41 crc kubenswrapper[4810]: I1008 06:50:41.970310 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/372bbd1e-a3e6-4f87-b824-4ee000006789-kube-api-access-sq8jv" (OuterVolumeSpecName: "kube-api-access-sq8jv") pod "372bbd1e-a3e6-4f87-b824-4ee000006789" (UID: "372bbd1e-a3e6-4f87-b824-4ee000006789"). InnerVolumeSpecName "kube-api-access-sq8jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.055934 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq8jv\" (UniqueName: \"kubernetes.io/projected/372bbd1e-a3e6-4f87-b824-4ee000006789-kube-api-access-sq8jv\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.202247 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerStarted","Data":"2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29"} Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.221627 4810 generic.go:334] "Generic (PLEG): container finished" podID="372bbd1e-a3e6-4f87-b824-4ee000006789" containerID="d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de" exitCode=2 Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.221872 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.222764 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"372bbd1e-a3e6-4f87-b824-4ee000006789","Type":"ContainerDied","Data":"d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de"} Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.222829 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"372bbd1e-a3e6-4f87-b824-4ee000006789","Type":"ContainerDied","Data":"c372558a4fa209c54d27868c155f6218c15707f4fd3c00ab75058d013de4576b"} Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.222852 4810 scope.go:117] "RemoveContainer" containerID="d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.236767 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-77482" event={"ID":"f7375a67-1872-4a18-a8b6-04cc8a6f06b8","Type":"ContainerDied","Data":"9241825651b89e0de120cb13a697f373ba57f94026cdad88f2152713fec8f376"} Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.236020 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7375a67-1872-4a18-a8b6-04cc8a6f06b8" containerID="9241825651b89e0de120cb13a697f373ba57f94026cdad88f2152713fec8f376" exitCode=0 Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.258628 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.276515 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.301041 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:50:42 crc kubenswrapper[4810]: E1008 06:50:42.301613 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="372bbd1e-a3e6-4f87-b824-4ee000006789" containerName="kube-state-metrics" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.301637 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="372bbd1e-a3e6-4f87-b824-4ee000006789" containerName="kube-state-metrics" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.301869 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="372bbd1e-a3e6-4f87-b824-4ee000006789" containerName="kube-state-metrics" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.302661 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.305667 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.310310 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.329461 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.341482 4810 scope.go:117] "RemoveContainer" containerID="d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de" Oct 08 06:50:42 crc kubenswrapper[4810]: E1008 06:50:42.345587 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de\": container with ID starting with d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de not found: ID does not exist" containerID="d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.345682 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de"} err="failed to get container status \"d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de\": rpc error: code = NotFound desc = could not find container \"d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de\": container with ID starting with d6393254d848e21f51e4ff00acc2f0e164479ee27985d2d30f284f2537d639de not found: ID does not exist" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.469502 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6942w\" (UniqueName: \"kubernetes.io/projected/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-api-access-6942w\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.469603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.469633 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.469686 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.571348 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.571795 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.571875 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.572010 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6942w\" (UniqueName: \"kubernetes.io/projected/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-api-access-6942w\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.587015 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.592220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.594559 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.604059 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6942w\" (UniqueName: \"kubernetes.io/projected/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-api-access-6942w\") pod \"kube-state-metrics-0\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " pod="openstack/kube-state-metrics-0" Oct 08 06:50:42 crc kubenswrapper[4810]: I1008 06:50:42.676383 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.167115 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.249900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce","Type":"ContainerStarted","Data":"72d60088be3097409ba2de5a4161245156bc48cdfc3bb9b38d7ae48729a5d505"} Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.254679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerStarted","Data":"48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0"} Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.685691 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.708911 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.800263 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-scripts\") pod \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.800338 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lwd2\" (UniqueName: \"kubernetes.io/projected/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-kube-api-access-4lwd2\") pod \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.800383 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-config-data\") pod \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.800549 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-combined-ca-bundle\") pod \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\" (UID: \"f7375a67-1872-4a18-a8b6-04cc8a6f06b8\") " Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.806050 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-kube-api-access-4lwd2" (OuterVolumeSpecName: "kube-api-access-4lwd2") pod "f7375a67-1872-4a18-a8b6-04cc8a6f06b8" (UID: "f7375a67-1872-4a18-a8b6-04cc8a6f06b8"). InnerVolumeSpecName "kube-api-access-4lwd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.806308 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-scripts" (OuterVolumeSpecName: "scripts") pod "f7375a67-1872-4a18-a8b6-04cc8a6f06b8" (UID: "f7375a67-1872-4a18-a8b6-04cc8a6f06b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.835702 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7375a67-1872-4a18-a8b6-04cc8a6f06b8" (UID: "f7375a67-1872-4a18-a8b6-04cc8a6f06b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.839602 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-config-data" (OuterVolumeSpecName: "config-data") pod "f7375a67-1872-4a18-a8b6-04cc8a6f06b8" (UID: "f7375a67-1872-4a18-a8b6-04cc8a6f06b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.902901 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.902953 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lwd2\" (UniqueName: \"kubernetes.io/projected/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-kube-api-access-4lwd2\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.903048 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:43 crc kubenswrapper[4810]: I1008 06:50:43.903059 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7375a67-1872-4a18-a8b6-04cc8a6f06b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.087892 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="372bbd1e-a3e6-4f87-b824-4ee000006789" path="/var/lib/kubelet/pods/372bbd1e-a3e6-4f87-b824-4ee000006789/volumes" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.266681 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-77482" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.266678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-77482" event={"ID":"f7375a67-1872-4a18-a8b6-04cc8a6f06b8","Type":"ContainerDied","Data":"ad52c257e232397edfe6378df53970737924ec4d54b8d71a32048766b50ad6b2"} Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.266916 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad52c257e232397edfe6378df53970737924ec4d54b8d71a32048766b50ad6b2" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.271083 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce","Type":"ContainerStarted","Data":"b767fec1d2cd5bc586cb8fb7ea2e613b95c815f5d05473a7605477f1e81ed54a"} Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.271404 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.275409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerStarted","Data":"0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640"} Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.275648 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-central-agent" containerID="cri-o://b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47" gracePeriod=30 Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.276063 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.276175 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="proxy-httpd" containerID="cri-o://0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640" gracePeriod=30 Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.276281 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="sg-core" containerID="cri-o://48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0" gracePeriod=30 Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.276377 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-notification-agent" containerID="cri-o://2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29" gracePeriod=30 Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.304700 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.945987618 podStartE2EDuration="2.30468113s" podCreationTimestamp="2025-10-08 06:50:42 +0000 UTC" firstStartedPulling="2025-10-08 06:50:43.165711249 +0000 UTC m=+1145.800150989" lastFinishedPulling="2025-10-08 06:50:43.524404761 +0000 UTC m=+1146.158844501" observedRunningTime="2025-10-08 06:50:44.294802366 +0000 UTC m=+1146.929242106" watchObservedRunningTime="2025-10-08 06:50:44.30468113 +0000 UTC m=+1146.939120870" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.336812 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.127356974 podStartE2EDuration="5.336785751s" podCreationTimestamp="2025-10-08 06:50:39 +0000 UTC" firstStartedPulling="2025-10-08 06:50:40.403224942 +0000 UTC m=+1143.037664682" lastFinishedPulling="2025-10-08 06:50:43.612653719 +0000 UTC m=+1146.247093459" observedRunningTime="2025-10-08 06:50:44.322401392 +0000 UTC m=+1146.956841132" watchObservedRunningTime="2025-10-08 06:50:44.336785751 +0000 UTC m=+1146.971225491" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.397707 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 06:50:44 crc kubenswrapper[4810]: E1008 06:50:44.398219 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7375a67-1872-4a18-a8b6-04cc8a6f06b8" containerName="nova-cell0-conductor-db-sync" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.398235 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7375a67-1872-4a18-a8b6-04cc8a6f06b8" containerName="nova-cell0-conductor-db-sync" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.398439 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7375a67-1872-4a18-a8b6-04cc8a6f06b8" containerName="nova-cell0-conductor-db-sync" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.399214 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.402390 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vtvjz" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.402398 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.413789 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.519624 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.519693 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfm8f\" (UniqueName: \"kubernetes.io/projected/ebbf3e4c-75db-4d41-8220-4b5c687d815f-kube-api-access-mfm8f\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.519728 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.621445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfm8f\" (UniqueName: \"kubernetes.io/projected/ebbf3e4c-75db-4d41-8220-4b5c687d815f-kube-api-access-mfm8f\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.622606 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.623538 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.627838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.638188 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.646077 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfm8f\" (UniqueName: \"kubernetes.io/projected/ebbf3e4c-75db-4d41-8220-4b5c687d815f-kube-api-access-mfm8f\") pod \"nova-cell0-conductor-0\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:44 crc kubenswrapper[4810]: I1008 06:50:44.769767 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:45 crc kubenswrapper[4810]: I1008 06:50:45.292090 4810 generic.go:334] "Generic (PLEG): container finished" podID="e06a2355-ff17-4cb4-bd37-63387901529e" containerID="0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640" exitCode=0 Oct 08 06:50:45 crc kubenswrapper[4810]: I1008 06:50:45.292545 4810 generic.go:334] "Generic (PLEG): container finished" podID="e06a2355-ff17-4cb4-bd37-63387901529e" containerID="48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0" exitCode=2 Oct 08 06:50:45 crc kubenswrapper[4810]: I1008 06:50:45.292557 4810 generic.go:334] "Generic (PLEG): container finished" podID="e06a2355-ff17-4cb4-bd37-63387901529e" containerID="2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29" exitCode=0 Oct 08 06:50:45 crc kubenswrapper[4810]: I1008 06:50:45.292144 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerDied","Data":"0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640"} Oct 08 06:50:45 crc kubenswrapper[4810]: I1008 06:50:45.292657 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerDied","Data":"48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0"} Oct 08 06:50:45 crc kubenswrapper[4810]: I1008 06:50:45.292674 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerDied","Data":"2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29"} Oct 08 06:50:45 crc kubenswrapper[4810]: I1008 06:50:45.299371 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 06:50:45 crc kubenswrapper[4810]: W1008 06:50:45.306650 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebbf3e4c_75db_4d41_8220_4b5c687d815f.slice/crio-0ecac1e8f1b2d9b982fb0729466a98a3325ac5db4d13bdf99ea89cd334ea098f WatchSource:0}: Error finding container 0ecac1e8f1b2d9b982fb0729466a98a3325ac5db4d13bdf99ea89cd334ea098f: Status 404 returned error can't find the container with id 0ecac1e8f1b2d9b982fb0729466a98a3325ac5db4d13bdf99ea89cd334ea098f Oct 08 06:50:46 crc kubenswrapper[4810]: I1008 06:50:46.307293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebbf3e4c-75db-4d41-8220-4b5c687d815f","Type":"ContainerStarted","Data":"0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe"} Oct 08 06:50:46 crc kubenswrapper[4810]: I1008 06:50:46.307951 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:46 crc kubenswrapper[4810]: I1008 06:50:46.307984 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebbf3e4c-75db-4d41-8220-4b5c687d815f","Type":"ContainerStarted","Data":"0ecac1e8f1b2d9b982fb0729466a98a3325ac5db4d13bdf99ea89cd334ea098f"} Oct 08 06:50:46 crc kubenswrapper[4810]: I1008 06:50:46.338696 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.338671004 podStartE2EDuration="2.338671004s" podCreationTimestamp="2025-10-08 06:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:46.33273647 +0000 UTC m=+1148.967176220" watchObservedRunningTime="2025-10-08 06:50:46.338671004 +0000 UTC m=+1148.973110744" Oct 08 06:50:48 crc kubenswrapper[4810]: I1008 06:50:48.970508 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075122 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4mks\" (UniqueName: \"kubernetes.io/projected/e06a2355-ff17-4cb4-bd37-63387901529e-kube-api-access-v4mks\") pod \"e06a2355-ff17-4cb4-bd37-63387901529e\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075223 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-run-httpd\") pod \"e06a2355-ff17-4cb4-bd37-63387901529e\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075272 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-combined-ca-bundle\") pod \"e06a2355-ff17-4cb4-bd37-63387901529e\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075290 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-sg-core-conf-yaml\") pod \"e06a2355-ff17-4cb4-bd37-63387901529e\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075306 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-config-data\") pod \"e06a2355-ff17-4cb4-bd37-63387901529e\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-scripts\") pod \"e06a2355-ff17-4cb4-bd37-63387901529e\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075499 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-log-httpd\") pod \"e06a2355-ff17-4cb4-bd37-63387901529e\" (UID: \"e06a2355-ff17-4cb4-bd37-63387901529e\") " Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.075618 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e06a2355-ff17-4cb4-bd37-63387901529e" (UID: "e06a2355-ff17-4cb4-bd37-63387901529e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.076380 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e06a2355-ff17-4cb4-bd37-63387901529e" (UID: "e06a2355-ff17-4cb4-bd37-63387901529e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.076862 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.076884 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e06a2355-ff17-4cb4-bd37-63387901529e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.081178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e06a2355-ff17-4cb4-bd37-63387901529e-kube-api-access-v4mks" (OuterVolumeSpecName: "kube-api-access-v4mks") pod "e06a2355-ff17-4cb4-bd37-63387901529e" (UID: "e06a2355-ff17-4cb4-bd37-63387901529e"). InnerVolumeSpecName "kube-api-access-v4mks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.081416 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-scripts" (OuterVolumeSpecName: "scripts") pod "e06a2355-ff17-4cb4-bd37-63387901529e" (UID: "e06a2355-ff17-4cb4-bd37-63387901529e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.105298 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e06a2355-ff17-4cb4-bd37-63387901529e" (UID: "e06a2355-ff17-4cb4-bd37-63387901529e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.152368 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e06a2355-ff17-4cb4-bd37-63387901529e" (UID: "e06a2355-ff17-4cb4-bd37-63387901529e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.178447 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.178477 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.178486 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.178495 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4mks\" (UniqueName: \"kubernetes.io/projected/e06a2355-ff17-4cb4-bd37-63387901529e-kube-api-access-v4mks\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.184277 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-config-data" (OuterVolumeSpecName: "config-data") pod "e06a2355-ff17-4cb4-bd37-63387901529e" (UID: "e06a2355-ff17-4cb4-bd37-63387901529e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.279647 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e06a2355-ff17-4cb4-bd37-63387901529e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.382319 4810 generic.go:334] "Generic (PLEG): container finished" podID="e06a2355-ff17-4cb4-bd37-63387901529e" containerID="b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47" exitCode=0 Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.382366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerDied","Data":"b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47"} Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.382396 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e06a2355-ff17-4cb4-bd37-63387901529e","Type":"ContainerDied","Data":"cff6de5e294ab4401219b8934a4ba3a4f2024b615010017771ec09573b1fcabd"} Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.382413 4810 scope.go:117] "RemoveContainer" containerID="0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.382752 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.413671 4810 scope.go:117] "RemoveContainer" containerID="48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.424312 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.431794 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.431855 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.431912 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.432961 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe33a8b708a76b97f4ea809aa4cdda57b3d8c666bcf720b830743227cfcf88de"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.433090 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://fe33a8b708a76b97f4ea809aa4cdda57b3d8c666bcf720b830743227cfcf88de" gracePeriod=600 Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.440069 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.450199 4810 scope.go:117] "RemoveContainer" containerID="2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.457960 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.459460 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-notification-agent" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.459497 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-notification-agent" Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.459524 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="sg-core" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.459536 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="sg-core" Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.459564 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-central-agent" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.459576 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-central-agent" Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.459602 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="proxy-httpd" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.459614 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="proxy-httpd" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.459997 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-notification-agent" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.460027 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="sg-core" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.460063 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="ceilometer-central-agent" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.460083 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" containerName="proxy-httpd" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.464355 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.471300 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.471343 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.473535 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484198 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484271 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-log-httpd\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484324 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484352 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-scripts\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484382 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzhcg\" (UniqueName: \"kubernetes.io/projected/c8854e99-28ee-4a93-b755-f362a8bf1c51-kube-api-access-kzhcg\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484504 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-run-httpd\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.484611 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-config-data\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.488076 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.495990 4810 scope.go:117] "RemoveContainer" containerID="b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.525349 4810 scope.go:117] "RemoveContainer" containerID="0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640" Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.525902 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640\": container with ID starting with 0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640 not found: ID does not exist" containerID="0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.526002 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640"} err="failed to get container status \"0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640\": rpc error: code = NotFound desc = could not find container \"0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640\": container with ID starting with 0566363f7107a6c5bf899493f58c640cd2a2f9b32ace38b1db72be399d38a640 not found: ID does not exist" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.526043 4810 scope.go:117] "RemoveContainer" containerID="48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0" Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.526588 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0\": container with ID starting with 48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0 not found: ID does not exist" containerID="48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.526666 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0"} err="failed to get container status \"48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0\": rpc error: code = NotFound desc = could not find container \"48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0\": container with ID starting with 48eb530341759ae47cf8673191bf316f7346c84c3f0bb689625c5c11b94e99a0 not found: ID does not exist" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.526693 4810 scope.go:117] "RemoveContainer" containerID="2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29" Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.527011 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29\": container with ID starting with 2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29 not found: ID does not exist" containerID="2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.527067 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29"} err="failed to get container status \"2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29\": rpc error: code = NotFound desc = could not find container \"2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29\": container with ID starting with 2bf7cc8efa46aa60c7e7189a429c659e0561fa821412f592c10154ddd1833b29 not found: ID does not exist" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.527100 4810 scope.go:117] "RemoveContainer" containerID="b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47" Oct 08 06:50:49 crc kubenswrapper[4810]: E1008 06:50:49.527401 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47\": container with ID starting with b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47 not found: ID does not exist" containerID="b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.527434 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47"} err="failed to get container status \"b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47\": rpc error: code = NotFound desc = could not find container \"b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47\": container with ID starting with b2474832180cdd2515e242726d420ad6e8c9fe6b46200f7c2273eeb1e9e94c47 not found: ID does not exist" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.587304 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-config-data\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.587774 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.587833 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-log-httpd\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.587885 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.587920 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-scripts\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.587949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.588023 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzhcg\" (UniqueName: \"kubernetes.io/projected/c8854e99-28ee-4a93-b755-f362a8bf1c51-kube-api-access-kzhcg\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.588054 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-run-httpd\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.588306 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-log-httpd\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.588693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-run-httpd\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.593053 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.593209 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-config-data\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.593331 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.594471 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.599154 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-scripts\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.604594 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzhcg\" (UniqueName: \"kubernetes.io/projected/c8854e99-28ee-4a93-b755-f362a8bf1c51-kube-api-access-kzhcg\") pod \"ceilometer-0\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " pod="openstack/ceilometer-0" Oct 08 06:50:49 crc kubenswrapper[4810]: I1008 06:50:49.799258 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:50:50 crc kubenswrapper[4810]: I1008 06:50:50.103055 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e06a2355-ff17-4cb4-bd37-63387901529e" path="/var/lib/kubelet/pods/e06a2355-ff17-4cb4-bd37-63387901529e/volumes" Oct 08 06:50:50 crc kubenswrapper[4810]: I1008 06:50:50.267838 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:50:50 crc kubenswrapper[4810]: I1008 06:50:50.394425 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="fe33a8b708a76b97f4ea809aa4cdda57b3d8c666bcf720b830743227cfcf88de" exitCode=0 Oct 08 06:50:50 crc kubenswrapper[4810]: I1008 06:50:50.394504 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"fe33a8b708a76b97f4ea809aa4cdda57b3d8c666bcf720b830743227cfcf88de"} Oct 08 06:50:50 crc kubenswrapper[4810]: I1008 06:50:50.394543 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"1fd9fd64ac269c2eaf81957ab40e460fd424d99aa24609f1272b49a186f84120"} Oct 08 06:50:50 crc kubenswrapper[4810]: I1008 06:50:50.394562 4810 scope.go:117] "RemoveContainer" containerID="bf6102000b45e1894934eca5fb9ef2a49eefef58f2ba5b5554021b8f1388a9d9" Oct 08 06:50:50 crc kubenswrapper[4810]: I1008 06:50:50.397756 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerStarted","Data":"17d7a0799428358e0244666f47bb8077ae4eddc345c36b31374852b3df9f3ff0"} Oct 08 06:50:51 crc kubenswrapper[4810]: I1008 06:50:51.422470 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerStarted","Data":"9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451"} Oct 08 06:50:52 crc kubenswrapper[4810]: I1008 06:50:52.433387 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerStarted","Data":"54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b"} Oct 08 06:50:52 crc kubenswrapper[4810]: I1008 06:50:52.689428 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 06:50:53 crc kubenswrapper[4810]: I1008 06:50:53.445411 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerStarted","Data":"569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec"} Oct 08 06:50:54 crc kubenswrapper[4810]: I1008 06:50:54.461914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerStarted","Data":"c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd"} Oct 08 06:50:54 crc kubenswrapper[4810]: I1008 06:50:54.462526 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 06:50:54 crc kubenswrapper[4810]: I1008 06:50:54.503624 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.060773593 podStartE2EDuration="5.503482803s" podCreationTimestamp="2025-10-08 06:50:49 +0000 UTC" firstStartedPulling="2025-10-08 06:50:50.274134307 +0000 UTC m=+1152.908574057" lastFinishedPulling="2025-10-08 06:50:53.716843527 +0000 UTC m=+1156.351283267" observedRunningTime="2025-10-08 06:50:54.483718445 +0000 UTC m=+1157.118158185" watchObservedRunningTime="2025-10-08 06:50:54.503482803 +0000 UTC m=+1157.137922563" Oct 08 06:50:54 crc kubenswrapper[4810]: I1008 06:50:54.813769 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.282710 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-trvz6"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.284804 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.289561 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.289811 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.307537 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-trvz6"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.419867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-config-data\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.419983 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-scripts\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.420022 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmjxl\" (UniqueName: \"kubernetes.io/projected/d318fcf9-d106-40cd-8ff9-d7e5abf43374-kube-api-access-dmjxl\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.420119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.450007 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.452378 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.457900 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.460795 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.521425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-scripts\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.521486 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmjxl\" (UniqueName: \"kubernetes.io/projected/d318fcf9-d106-40cd-8ff9-d7e5abf43374-kube-api-access-dmjxl\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.521555 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.521629 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-config-data\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.530945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.534276 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-scripts\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.536051 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-config-data\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.552049 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmjxl\" (UniqueName: \"kubernetes.io/projected/d318fcf9-d106-40cd-8ff9-d7e5abf43374-kube-api-access-dmjxl\") pod \"nova-cell0-cell-mapping-trvz6\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.558445 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.562866 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.575258 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.592896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.616154 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.625744 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-config-data\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.625814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-logs\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.625887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.626064 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgzr2\" (UniqueName: \"kubernetes.io/projected/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-kube-api-access-jgzr2\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.695588 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.697201 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.720842 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.727979 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-config-data\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.728090 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgzr2\" (UniqueName: \"kubernetes.io/projected/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-kube-api-access-jgzr2\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.728134 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcz9f\" (UniqueName: \"kubernetes.io/projected/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-kube-api-access-wcz9f\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.728220 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.728455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-config-data\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.728492 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-logs\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.728589 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-logs\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.728622 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.734727 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-logs\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.742135 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.747115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-config-data\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.753748 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.764019 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.775624 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.776292 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgzr2\" (UniqueName: \"kubernetes.io/projected/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-kube-api-access-jgzr2\") pod \"nova-api-0\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.791309 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.791906 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.820727 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.833279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.833348 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txfdl\" (UniqueName: \"kubernetes.io/projected/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-kube-api-access-txfdl\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.833416 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-config-data\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.833497 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcz9f\" (UniqueName: \"kubernetes.io/projected/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-kube-api-access-wcz9f\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.833558 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.833741 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.833795 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-logs\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.834280 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-logs\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.839790 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.851504 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-config-data\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.857717 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcz9f\" (UniqueName: \"kubernetes.io/projected/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-kube-api-access-wcz9f\") pod \"nova-metadata-0\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " pod="openstack/nova-metadata-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.868096 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6ffc974fdf-6dtwt"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.869846 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.877095 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffc974fdf-6dtwt"] Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.936112 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.936222 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.936278 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.936302 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6w5z\" (UniqueName: \"kubernetes.io/projected/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-kube-api-access-w6w5z\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.936324 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txfdl\" (UniqueName: \"kubernetes.io/projected/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-kube-api-access-txfdl\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.936360 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-config-data\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.940476 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.940958 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:55 crc kubenswrapper[4810]: I1008 06:50:55.968566 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txfdl\" (UniqueName: \"kubernetes.io/projected/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-kube-api-access-txfdl\") pod \"nova-cell1-novncproxy-0\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.028194 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041527 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6w5z\" (UniqueName: \"kubernetes.io/projected/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-kube-api-access-w6w5z\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041573 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041614 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041637 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-config-data\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041671 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzvlv\" (UniqueName: \"kubernetes.io/projected/89c96d63-b16e-40bb-9acd-a60a85cd8af2-kube-api-access-bzvlv\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041705 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-svc\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041734 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041800 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-config\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.041820 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-swift-storage-0\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.049679 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.053383 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-config-data\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.066706 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6w5z\" (UniqueName: \"kubernetes.io/projected/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-kube-api-access-w6w5z\") pod \"nova-scheduler-0\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " pod="openstack/nova-scheduler-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.128828 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.129614 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.144090 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-config\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.144131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-swift-storage-0\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.144185 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.144216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.144247 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzvlv\" (UniqueName: \"kubernetes.io/projected/89c96d63-b16e-40bb-9acd-a60a85cd8af2-kube-api-access-bzvlv\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.144279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-svc\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.145516 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-svc\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.145714 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.146121 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.146846 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-config\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.147369 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-swift-storage-0\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.172003 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzvlv\" (UniqueName: \"kubernetes.io/projected/89c96d63-b16e-40bb-9acd-a60a85cd8af2-kube-api-access-bzvlv\") pod \"dnsmasq-dns-6ffc974fdf-6dtwt\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.205193 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.375055 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-trvz6"] Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.408407 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.501761 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf","Type":"ContainerStarted","Data":"fbbac990da7329a33c1382cb9d1854733a890928a8001af02f1586417662c368"} Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.504076 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-trvz6" event={"ID":"d318fcf9-d106-40cd-8ff9-d7e5abf43374","Type":"ContainerStarted","Data":"e86dcd85db99a4e923e813875835e1b595afd078958500d017f42544e047de46"} Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.581241 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.769020 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.866683 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffc974fdf-6dtwt"] Oct 08 06:50:56 crc kubenswrapper[4810]: W1008 06:50:56.881151 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89c96d63_b16e_40bb_9acd_a60a85cd8af2.slice/crio-b8e47bdb3ef319497fa8bc576caf7b6dddbf1acae2053881413906fe60bb8914 WatchSource:0}: Error finding container b8e47bdb3ef319497fa8bc576caf7b6dddbf1acae2053881413906fe60bb8914: Status 404 returned error can't find the container with id b8e47bdb3ef319497fa8bc576caf7b6dddbf1acae2053881413906fe60bb8914 Oct 08 06:50:56 crc kubenswrapper[4810]: I1008 06:50:56.957696 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.043089 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ncgvz"] Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.044865 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.048453 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.050459 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.063234 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ncgvz"] Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.195237 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.195596 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9542r\" (UniqueName: \"kubernetes.io/projected/40eae6ac-32d1-406c-b971-04621dcfaeb3-kube-api-access-9542r\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.195668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-scripts\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.195713 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-config-data\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.297444 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9542r\" (UniqueName: \"kubernetes.io/projected/40eae6ac-32d1-406c-b971-04621dcfaeb3-kube-api-access-9542r\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.297507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-scripts\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.297533 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-config-data\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.297597 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.305590 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-scripts\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.307045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-config-data\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.310263 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.317421 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9542r\" (UniqueName: \"kubernetes.io/projected/40eae6ac-32d1-406c-b971-04621dcfaeb3-kube-api-access-9542r\") pod \"nova-cell1-conductor-db-sync-ncgvz\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.403542 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.525721 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9","Type":"ContainerStarted","Data":"977432afab6d0cfd25f035fbb91479bfbb3b320b5d0e56f3aea105a1e9bba45a"} Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.549032 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d","Type":"ContainerStarted","Data":"dd4746bc63f1d079c5934ee674303867f70a157b5b99656ca8c54b67302944ea"} Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.551626 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1adfe378-7e5d-4274-a5c5-0ab624c84bcf","Type":"ContainerStarted","Data":"49c4adbec3c13d78a266ea6df1dab9a03eec80e30eb0b8f12f7d2fd46e9dec27"} Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.554435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-trvz6" event={"ID":"d318fcf9-d106-40cd-8ff9-d7e5abf43374","Type":"ContainerStarted","Data":"21055dd63bbf423089e26821be7adc73c66cfc6acb4e71a03dd594f0a5233b3c"} Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.560082 4810 generic.go:334] "Generic (PLEG): container finished" podID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerID="fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751" exitCode=0 Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.560135 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" event={"ID":"89c96d63-b16e-40bb-9acd-a60a85cd8af2","Type":"ContainerDied","Data":"fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751"} Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.560173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" event={"ID":"89c96d63-b16e-40bb-9acd-a60a85cd8af2","Type":"ContainerStarted","Data":"b8e47bdb3ef319497fa8bc576caf7b6dddbf1acae2053881413906fe60bb8914"} Oct 08 06:50:57 crc kubenswrapper[4810]: I1008 06:50:57.599143 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-trvz6" podStartSLOduration=2.599122854 podStartE2EDuration="2.599122854s" podCreationTimestamp="2025-10-08 06:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:57.585144566 +0000 UTC m=+1160.219584306" watchObservedRunningTime="2025-10-08 06:50:57.599122854 +0000 UTC m=+1160.233562594" Oct 08 06:50:58 crc kubenswrapper[4810]: I1008 06:50:58.054030 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ncgvz"] Oct 08 06:50:58 crc kubenswrapper[4810]: W1008 06:50:58.073509 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40eae6ac_32d1_406c_b971_04621dcfaeb3.slice/crio-e62c999facf54b1628887a2ba311e0741adab865d48d7cd34b6669bc4e0f14ad WatchSource:0}: Error finding container e62c999facf54b1628887a2ba311e0741adab865d48d7cd34b6669bc4e0f14ad: Status 404 returned error can't find the container with id e62c999facf54b1628887a2ba311e0741adab865d48d7cd34b6669bc4e0f14ad Oct 08 06:50:58 crc kubenswrapper[4810]: I1008 06:50:58.576263 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" event={"ID":"40eae6ac-32d1-406c-b971-04621dcfaeb3","Type":"ContainerStarted","Data":"623a3988e5bd08655720c75adbeab397cb0959190125b45cf629d8f3da4bdd2d"} Oct 08 06:50:58 crc kubenswrapper[4810]: I1008 06:50:58.576822 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" event={"ID":"40eae6ac-32d1-406c-b971-04621dcfaeb3","Type":"ContainerStarted","Data":"e62c999facf54b1628887a2ba311e0741adab865d48d7cd34b6669bc4e0f14ad"} Oct 08 06:50:58 crc kubenswrapper[4810]: I1008 06:50:58.587792 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" event={"ID":"89c96d63-b16e-40bb-9acd-a60a85cd8af2","Type":"ContainerStarted","Data":"94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339"} Oct 08 06:50:58 crc kubenswrapper[4810]: I1008 06:50:58.588063 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:50:58 crc kubenswrapper[4810]: I1008 06:50:58.600603 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" podStartSLOduration=1.6005846 podStartE2EDuration="1.6005846s" podCreationTimestamp="2025-10-08 06:50:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:58.596071395 +0000 UTC m=+1161.230511135" watchObservedRunningTime="2025-10-08 06:50:58.6005846 +0000 UTC m=+1161.235024330" Oct 08 06:50:58 crc kubenswrapper[4810]: I1008 06:50:58.621781 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" podStartSLOduration=3.621754468 podStartE2EDuration="3.621754468s" podCreationTimestamp="2025-10-08 06:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:50:58.620172063 +0000 UTC m=+1161.254611823" watchObservedRunningTime="2025-10-08 06:50:58.621754468 +0000 UTC m=+1161.256194208" Oct 08 06:50:59 crc kubenswrapper[4810]: I1008 06:50:59.514041 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:50:59 crc kubenswrapper[4810]: I1008 06:50:59.527798 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.634185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1adfe378-7e5d-4274-a5c5-0ab624c84bcf","Type":"ContainerStarted","Data":"2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763"} Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.636068 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1adfe378-7e5d-4274-a5c5-0ab624c84bcf" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763" gracePeriod=30 Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.649651 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf","Type":"ContainerStarted","Data":"b1c8620d6969a132a8688a38415693d0c85acae6386ee9cee039b3d30043d082"} Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.649720 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf","Type":"ContainerStarted","Data":"8a8635f25b51d99c85a053c1c6540e7eace15f2f55710edcd7fbf48a404e5395"} Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.652807 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9","Type":"ContainerStarted","Data":"83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0"} Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.656302 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d","Type":"ContainerStarted","Data":"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37"} Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.656373 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d","Type":"ContainerStarted","Data":"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b"} Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.656371 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-log" containerID="cri-o://0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b" gracePeriod=30 Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.656381 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-metadata" containerID="cri-o://2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37" gracePeriod=30 Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.670292 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.927940879 podStartE2EDuration="6.670260861s" podCreationTimestamp="2025-10-08 06:50:55 +0000 UTC" firstStartedPulling="2025-10-08 06:50:56.766206585 +0000 UTC m=+1159.400646325" lastFinishedPulling="2025-10-08 06:51:00.508526557 +0000 UTC m=+1163.142966307" observedRunningTime="2025-10-08 06:51:01.663325319 +0000 UTC m=+1164.297765069" watchObservedRunningTime="2025-10-08 06:51:01.670260861 +0000 UTC m=+1164.304700601" Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.712870 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.194351069 podStartE2EDuration="6.712841492s" podCreationTimestamp="2025-10-08 06:50:55 +0000 UTC" firstStartedPulling="2025-10-08 06:50:56.973026012 +0000 UTC m=+1159.607465752" lastFinishedPulling="2025-10-08 06:51:00.491516435 +0000 UTC m=+1163.125956175" observedRunningTime="2025-10-08 06:51:01.693736592 +0000 UTC m=+1164.328176332" watchObservedRunningTime="2025-10-08 06:51:01.712841492 +0000 UTC m=+1164.347281252" Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.725241 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.827934323 podStartE2EDuration="6.725220116s" podCreationTimestamp="2025-10-08 06:50:55 +0000 UTC" firstStartedPulling="2025-10-08 06:50:56.59486187 +0000 UTC m=+1159.229301610" lastFinishedPulling="2025-10-08 06:51:00.492147623 +0000 UTC m=+1163.126587403" observedRunningTime="2025-10-08 06:51:01.720908996 +0000 UTC m=+1164.355348736" watchObservedRunningTime="2025-10-08 06:51:01.725220116 +0000 UTC m=+1164.359659846" Oct 08 06:51:01 crc kubenswrapper[4810]: I1008 06:51:01.751881 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.654194292 podStartE2EDuration="6.751865115s" podCreationTimestamp="2025-10-08 06:50:55 +0000 UTC" firstStartedPulling="2025-10-08 06:50:56.408630583 +0000 UTC m=+1159.043070323" lastFinishedPulling="2025-10-08 06:51:00.506301396 +0000 UTC m=+1163.140741146" observedRunningTime="2025-10-08 06:51:01.750870467 +0000 UTC m=+1164.385310207" watchObservedRunningTime="2025-10-08 06:51:01.751865115 +0000 UTC m=+1164.386304855" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.297483 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.351558 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcz9f\" (UniqueName: \"kubernetes.io/projected/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-kube-api-access-wcz9f\") pod \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.352099 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-logs\") pod \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.352180 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-config-data\") pod \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.352408 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-combined-ca-bundle\") pod \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\" (UID: \"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d\") " Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.352743 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-logs" (OuterVolumeSpecName: "logs") pod "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" (UID: "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.356770 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.364793 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-kube-api-access-wcz9f" (OuterVolumeSpecName: "kube-api-access-wcz9f") pod "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" (UID: "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d"). InnerVolumeSpecName "kube-api-access-wcz9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.387664 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-config-data" (OuterVolumeSpecName: "config-data") pod "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" (UID: "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.407123 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" (UID: "3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.460147 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.460195 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcz9f\" (UniqueName: \"kubernetes.io/projected/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-kube-api-access-wcz9f\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.460214 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.671361 4810 generic.go:334] "Generic (PLEG): container finished" podID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerID="2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37" exitCode=0 Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.671409 4810 generic.go:334] "Generic (PLEG): container finished" podID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerID="0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b" exitCode=143 Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.671450 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d","Type":"ContainerDied","Data":"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37"} Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.671463 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.671499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d","Type":"ContainerDied","Data":"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b"} Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.671516 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d","Type":"ContainerDied","Data":"dd4746bc63f1d079c5934ee674303867f70a157b5b99656ca8c54b67302944ea"} Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.671547 4810 scope.go:117] "RemoveContainer" containerID="2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.713778 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.715358 4810 scope.go:117] "RemoveContainer" containerID="0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.724312 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.736399 4810 scope.go:117] "RemoveContainer" containerID="2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37" Oct 08 06:51:02 crc kubenswrapper[4810]: E1008 06:51:02.737029 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37\": container with ID starting with 2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37 not found: ID does not exist" containerID="2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.737075 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37"} err="failed to get container status \"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37\": rpc error: code = NotFound desc = could not find container \"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37\": container with ID starting with 2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37 not found: ID does not exist" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.737109 4810 scope.go:117] "RemoveContainer" containerID="0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b" Oct 08 06:51:02 crc kubenswrapper[4810]: E1008 06:51:02.738577 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b\": container with ID starting with 0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b not found: ID does not exist" containerID="0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.738634 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b"} err="failed to get container status \"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b\": rpc error: code = NotFound desc = could not find container \"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b\": container with ID starting with 0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b not found: ID does not exist" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.738679 4810 scope.go:117] "RemoveContainer" containerID="2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.739145 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37"} err="failed to get container status \"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37\": rpc error: code = NotFound desc = could not find container \"2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37\": container with ID starting with 2eb148937f61be1785bb279779d0518692fa6a9fc5b30cb65b30a02ffc80cb37 not found: ID does not exist" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.739220 4810 scope.go:117] "RemoveContainer" containerID="0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.739724 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b"} err="failed to get container status \"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b\": rpc error: code = NotFound desc = could not find container \"0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b\": container with ID starting with 0eaecf2446ea84902feaa2f753e84a65862d0da03514db8e01f206f1eb1d0b6b not found: ID does not exist" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.749705 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:02 crc kubenswrapper[4810]: E1008 06:51:02.750167 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-metadata" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.750186 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-metadata" Oct 08 06:51:02 crc kubenswrapper[4810]: E1008 06:51:02.750215 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-log" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.750225 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-log" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.750441 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-metadata" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.750460 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" containerName="nova-metadata-log" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.751514 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.755489 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.755677 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.766657 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-config-data\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.766798 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.766878 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.766998 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e481cf82-f798-4da5-8ffc-1a737209af70-logs\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.767107 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59sp6\" (UniqueName: \"kubernetes.io/projected/e481cf82-f798-4da5-8ffc-1a737209af70-kube-api-access-59sp6\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.784260 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.867985 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59sp6\" (UniqueName: \"kubernetes.io/projected/e481cf82-f798-4da5-8ffc-1a737209af70-kube-api-access-59sp6\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.868046 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-config-data\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.868096 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.868140 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.869024 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e481cf82-f798-4da5-8ffc-1a737209af70-logs\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.869397 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e481cf82-f798-4da5-8ffc-1a737209af70-logs\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.875742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.880761 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.888357 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-config-data\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:02 crc kubenswrapper[4810]: I1008 06:51:02.900508 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59sp6\" (UniqueName: \"kubernetes.io/projected/e481cf82-f798-4da5-8ffc-1a737209af70-kube-api-access-59sp6\") pod \"nova-metadata-0\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " pod="openstack/nova-metadata-0" Oct 08 06:51:03 crc kubenswrapper[4810]: I1008 06:51:03.070722 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:03 crc kubenswrapper[4810]: I1008 06:51:03.547112 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:03 crc kubenswrapper[4810]: W1008 06:51:03.547286 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode481cf82_f798_4da5_8ffc_1a737209af70.slice/crio-ac360ebccadf6d249bbcba2784c097de44f32a6b3efe1f43d0127688d29a2bbf WatchSource:0}: Error finding container ac360ebccadf6d249bbcba2784c097de44f32a6b3efe1f43d0127688d29a2bbf: Status 404 returned error can't find the container with id ac360ebccadf6d249bbcba2784c097de44f32a6b3efe1f43d0127688d29a2bbf Oct 08 06:51:03 crc kubenswrapper[4810]: I1008 06:51:03.697586 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e481cf82-f798-4da5-8ffc-1a737209af70","Type":"ContainerStarted","Data":"ac360ebccadf6d249bbcba2784c097de44f32a6b3efe1f43d0127688d29a2bbf"} Oct 08 06:51:04 crc kubenswrapper[4810]: I1008 06:51:04.088541 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d" path="/var/lib/kubelet/pods/3ab07a7d-d6d5-4630-ba3e-5fab5e3bda2d/volumes" Oct 08 06:51:04 crc kubenswrapper[4810]: I1008 06:51:04.713388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e481cf82-f798-4da5-8ffc-1a737209af70","Type":"ContainerStarted","Data":"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e"} Oct 08 06:51:04 crc kubenswrapper[4810]: I1008 06:51:04.713451 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e481cf82-f798-4da5-8ffc-1a737209af70","Type":"ContainerStarted","Data":"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38"} Oct 08 06:51:04 crc kubenswrapper[4810]: I1008 06:51:04.753720 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.753696713 podStartE2EDuration="2.753696713s" podCreationTimestamp="2025-10-08 06:51:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:04.744998212 +0000 UTC m=+1167.379437962" watchObservedRunningTime="2025-10-08 06:51:04.753696713 +0000 UTC m=+1167.388136453" Oct 08 06:51:05 crc kubenswrapper[4810]: I1008 06:51:05.728418 4810 generic.go:334] "Generic (PLEG): container finished" podID="d318fcf9-d106-40cd-8ff9-d7e5abf43374" containerID="21055dd63bbf423089e26821be7adc73c66cfc6acb4e71a03dd594f0a5233b3c" exitCode=0 Oct 08 06:51:05 crc kubenswrapper[4810]: I1008 06:51:05.728572 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-trvz6" event={"ID":"d318fcf9-d106-40cd-8ff9-d7e5abf43374","Type":"ContainerDied","Data":"21055dd63bbf423089e26821be7adc73c66cfc6acb4e71a03dd594f0a5233b3c"} Oct 08 06:51:05 crc kubenswrapper[4810]: I1008 06:51:05.792926 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 06:51:05 crc kubenswrapper[4810]: I1008 06:51:05.792995 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.130183 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.130234 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.130249 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.175902 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.207201 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.292868 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bd785c49-86blx"] Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.293347 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84bd785c49-86blx" podUID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerName="dnsmasq-dns" containerID="cri-o://849c967153d2f165920e9dfd4ff8e9b6bd05ad3f465bda0698dd4e80edcf02a0" gracePeriod=10 Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.776097 4810 generic.go:334] "Generic (PLEG): container finished" podID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerID="849c967153d2f165920e9dfd4ff8e9b6bd05ad3f465bda0698dd4e80edcf02a0" exitCode=0 Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.776181 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bd785c49-86blx" event={"ID":"850b635b-9e46-4f19-b573-9dcde8ec4d24","Type":"ContainerDied","Data":"849c967153d2f165920e9dfd4ff8e9b6bd05ad3f465bda0698dd4e80edcf02a0"} Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.776239 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bd785c49-86blx" event={"ID":"850b635b-9e46-4f19-b573-9dcde8ec4d24","Type":"ContainerDied","Data":"7c7ad6b73903a4a9805471de20f27c4c392c49ef8d2c90254a13bd7794140719"} Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.776257 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c7ad6b73903a4a9805471de20f27c4c392c49ef8d2c90254a13bd7794140719" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.780136 4810 generic.go:334] "Generic (PLEG): container finished" podID="40eae6ac-32d1-406c-b971-04621dcfaeb3" containerID="623a3988e5bd08655720c75adbeab397cb0959190125b45cf629d8f3da4bdd2d" exitCode=0 Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.780239 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" event={"ID":"40eae6ac-32d1-406c-b971-04621dcfaeb3","Type":"ContainerDied","Data":"623a3988e5bd08655720c75adbeab397cb0959190125b45cf629d8f3da4bdd2d"} Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.822568 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.826922 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.857683 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-config\") pod \"850b635b-9e46-4f19-b573-9dcde8ec4d24\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.857770 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-svc\") pod \"850b635b-9e46-4f19-b573-9dcde8ec4d24\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.857815 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcf4t\" (UniqueName: \"kubernetes.io/projected/850b635b-9e46-4f19-b573-9dcde8ec4d24-kube-api-access-rcf4t\") pod \"850b635b-9e46-4f19-b573-9dcde8ec4d24\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.857835 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-sb\") pod \"850b635b-9e46-4f19-b573-9dcde8ec4d24\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.857952 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-nb\") pod \"850b635b-9e46-4f19-b573-9dcde8ec4d24\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.858039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-swift-storage-0\") pod \"850b635b-9e46-4f19-b573-9dcde8ec4d24\" (UID: \"850b635b-9e46-4f19-b573-9dcde8ec4d24\") " Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.871406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850b635b-9e46-4f19-b573-9dcde8ec4d24-kube-api-access-rcf4t" (OuterVolumeSpecName: "kube-api-access-rcf4t") pod "850b635b-9e46-4f19-b573-9dcde8ec4d24" (UID: "850b635b-9e46-4f19-b573-9dcde8ec4d24"). InnerVolumeSpecName "kube-api-access-rcf4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.878488 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.878508 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.960934 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcf4t\" (UniqueName: \"kubernetes.io/projected/850b635b-9e46-4f19-b573-9dcde8ec4d24-kube-api-access-rcf4t\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.960925 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-config" (OuterVolumeSpecName: "config") pod "850b635b-9e46-4f19-b573-9dcde8ec4d24" (UID: "850b635b-9e46-4f19-b573-9dcde8ec4d24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:06 crc kubenswrapper[4810]: I1008 06:51:06.964833 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "850b635b-9e46-4f19-b573-9dcde8ec4d24" (UID: "850b635b-9e46-4f19-b573-9dcde8ec4d24"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.004520 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "850b635b-9e46-4f19-b573-9dcde8ec4d24" (UID: "850b635b-9e46-4f19-b573-9dcde8ec4d24"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.035623 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "850b635b-9e46-4f19-b573-9dcde8ec4d24" (UID: "850b635b-9e46-4f19-b573-9dcde8ec4d24"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.051242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "850b635b-9e46-4f19-b573-9dcde8ec4d24" (UID: "850b635b-9e46-4f19-b573-9dcde8ec4d24"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.065156 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.065200 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.065211 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.065222 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.065232 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/850b635b-9e46-4f19-b573-9dcde8ec4d24-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.132607 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.167205 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-config-data\") pod \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.167578 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmjxl\" (UniqueName: \"kubernetes.io/projected/d318fcf9-d106-40cd-8ff9-d7e5abf43374-kube-api-access-dmjxl\") pod \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.168104 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-scripts\") pod \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.168283 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-combined-ca-bundle\") pod \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\" (UID: \"d318fcf9-d106-40cd-8ff9-d7e5abf43374\") " Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.179613 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d318fcf9-d106-40cd-8ff9-d7e5abf43374-kube-api-access-dmjxl" (OuterVolumeSpecName: "kube-api-access-dmjxl") pod "d318fcf9-d106-40cd-8ff9-d7e5abf43374" (UID: "d318fcf9-d106-40cd-8ff9-d7e5abf43374"). InnerVolumeSpecName "kube-api-access-dmjxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.179617 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-scripts" (OuterVolumeSpecName: "scripts") pod "d318fcf9-d106-40cd-8ff9-d7e5abf43374" (UID: "d318fcf9-d106-40cd-8ff9-d7e5abf43374"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.199652 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d318fcf9-d106-40cd-8ff9-d7e5abf43374" (UID: "d318fcf9-d106-40cd-8ff9-d7e5abf43374"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.199707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-config-data" (OuterVolumeSpecName: "config-data") pod "d318fcf9-d106-40cd-8ff9-d7e5abf43374" (UID: "d318fcf9-d106-40cd-8ff9-d7e5abf43374"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.271896 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.271935 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.271948 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmjxl\" (UniqueName: \"kubernetes.io/projected/d318fcf9-d106-40cd-8ff9-d7e5abf43374-kube-api-access-dmjxl\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.271959 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d318fcf9-d106-40cd-8ff9-d7e5abf43374-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.795879 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-trvz6" event={"ID":"d318fcf9-d106-40cd-8ff9-d7e5abf43374","Type":"ContainerDied","Data":"e86dcd85db99a4e923e813875835e1b595afd078958500d017f42544e047de46"} Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.796307 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e86dcd85db99a4e923e813875835e1b595afd078958500d017f42544e047de46" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.795944 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd785c49-86blx" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.795951 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-trvz6" Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.865670 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bd785c49-86blx"] Oct 08 06:51:07 crc kubenswrapper[4810]: I1008 06:51:07.875902 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84bd785c49-86blx"] Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.032980 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.033770 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-log" containerID="cri-o://8a8635f25b51d99c85a053c1c6540e7eace15f2f55710edcd7fbf48a404e5395" gracePeriod=30 Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.034814 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-api" containerID="cri-o://b1c8620d6969a132a8688a38415693d0c85acae6386ee9cee039b3d30043d082" gracePeriod=30 Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.071401 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.071827 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.120210 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850b635b-9e46-4f19-b573-9dcde8ec4d24" path="/var/lib/kubelet/pods/850b635b-9e46-4f19-b573-9dcde8ec4d24/volumes" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.121649 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.121783 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.250545 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.293218 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-combined-ca-bundle\") pod \"40eae6ac-32d1-406c-b971-04621dcfaeb3\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.293372 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9542r\" (UniqueName: \"kubernetes.io/projected/40eae6ac-32d1-406c-b971-04621dcfaeb3-kube-api-access-9542r\") pod \"40eae6ac-32d1-406c-b971-04621dcfaeb3\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.293473 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-scripts\") pod \"40eae6ac-32d1-406c-b971-04621dcfaeb3\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.293495 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-config-data\") pod \"40eae6ac-32d1-406c-b971-04621dcfaeb3\" (UID: \"40eae6ac-32d1-406c-b971-04621dcfaeb3\") " Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.299106 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40eae6ac-32d1-406c-b971-04621dcfaeb3-kube-api-access-9542r" (OuterVolumeSpecName: "kube-api-access-9542r") pod "40eae6ac-32d1-406c-b971-04621dcfaeb3" (UID: "40eae6ac-32d1-406c-b971-04621dcfaeb3"). InnerVolumeSpecName "kube-api-access-9542r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.300599 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-scripts" (OuterVolumeSpecName: "scripts") pod "40eae6ac-32d1-406c-b971-04621dcfaeb3" (UID: "40eae6ac-32d1-406c-b971-04621dcfaeb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:08 crc kubenswrapper[4810]: E1008 06:51:08.315607 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4b1bf70_6a1e_4c52_aa30_6b17c91397cf.slice/crio-conmon-8a8635f25b51d99c85a053c1c6540e7eace15f2f55710edcd7fbf48a404e5395.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4b1bf70_6a1e_4c52_aa30_6b17c91397cf.slice/crio-8a8635f25b51d99c85a053c1c6540e7eace15f2f55710edcd7fbf48a404e5395.scope\": RecentStats: unable to find data in memory cache]" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.323995 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-config-data" (OuterVolumeSpecName: "config-data") pod "40eae6ac-32d1-406c-b971-04621dcfaeb3" (UID: "40eae6ac-32d1-406c-b971-04621dcfaeb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.342181 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40eae6ac-32d1-406c-b971-04621dcfaeb3" (UID: "40eae6ac-32d1-406c-b971-04621dcfaeb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.396409 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9542r\" (UniqueName: \"kubernetes.io/projected/40eae6ac-32d1-406c-b971-04621dcfaeb3-kube-api-access-9542r\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.396445 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.396455 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.396466 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40eae6ac-32d1-406c-b971-04621dcfaeb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.817689 4810 generic.go:334] "Generic (PLEG): container finished" podID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerID="8a8635f25b51d99c85a053c1c6540e7eace15f2f55710edcd7fbf48a404e5395" exitCode=143 Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.818038 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf","Type":"ContainerDied","Data":"8a8635f25b51d99c85a053c1c6540e7eace15f2f55710edcd7fbf48a404e5395"} Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.822347 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" event={"ID":"40eae6ac-32d1-406c-b971-04621dcfaeb3","Type":"ContainerDied","Data":"e62c999facf54b1628887a2ba311e0741adab865d48d7cd34b6669bc4e0f14ad"} Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.822407 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e62c999facf54b1628887a2ba311e0741adab865d48d7cd34b6669bc4e0f14ad" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.822418 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ncgvz" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.822575 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" containerName="nova-scheduler-scheduler" containerID="cri-o://83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0" gracePeriod=30 Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.893806 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 06:51:08 crc kubenswrapper[4810]: E1008 06:51:08.894214 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d318fcf9-d106-40cd-8ff9-d7e5abf43374" containerName="nova-manage" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.894234 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d318fcf9-d106-40cd-8ff9-d7e5abf43374" containerName="nova-manage" Oct 08 06:51:08 crc kubenswrapper[4810]: E1008 06:51:08.894253 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerName="init" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.894262 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerName="init" Oct 08 06:51:08 crc kubenswrapper[4810]: E1008 06:51:08.894285 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40eae6ac-32d1-406c-b971-04621dcfaeb3" containerName="nova-cell1-conductor-db-sync" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.894291 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="40eae6ac-32d1-406c-b971-04621dcfaeb3" containerName="nova-cell1-conductor-db-sync" Oct 08 06:51:08 crc kubenswrapper[4810]: E1008 06:51:08.894331 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerName="dnsmasq-dns" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.894336 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerName="dnsmasq-dns" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.894525 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="850b635b-9e46-4f19-b573-9dcde8ec4d24" containerName="dnsmasq-dns" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.894545 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d318fcf9-d106-40cd-8ff9-d7e5abf43374" containerName="nova-manage" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.894558 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="40eae6ac-32d1-406c-b971-04621dcfaeb3" containerName="nova-cell1-conductor-db-sync" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.895259 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.897270 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 06:51:08 crc kubenswrapper[4810]: I1008 06:51:08.903401 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.009877 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.010005 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xp5b\" (UniqueName: \"kubernetes.io/projected/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-kube-api-access-2xp5b\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.010048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.111553 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.111673 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xp5b\" (UniqueName: \"kubernetes.io/projected/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-kube-api-access-2xp5b\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.111704 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.118010 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.128772 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.131088 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xp5b\" (UniqueName: \"kubernetes.io/projected/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-kube-api-access-2xp5b\") pod \"nova-cell1-conductor-0\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.214504 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.682432 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.858142 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b","Type":"ContainerStarted","Data":"b8dd244f52cb0ecb3525951046b786fa3edfb96ad12d542da4ee883c31710832"} Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.858368 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-log" containerID="cri-o://7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38" gracePeriod=30 Oct 08 06:51:09 crc kubenswrapper[4810]: I1008 06:51:09.858572 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-metadata" containerID="cri-o://e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e" gracePeriod=30 Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.469640 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.549495 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-nova-metadata-tls-certs\") pod \"e481cf82-f798-4da5-8ffc-1a737209af70\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.549560 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e481cf82-f798-4da5-8ffc-1a737209af70-logs\") pod \"e481cf82-f798-4da5-8ffc-1a737209af70\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.549664 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-config-data\") pod \"e481cf82-f798-4da5-8ffc-1a737209af70\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.549714 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-combined-ca-bundle\") pod \"e481cf82-f798-4da5-8ffc-1a737209af70\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.549976 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59sp6\" (UniqueName: \"kubernetes.io/projected/e481cf82-f798-4da5-8ffc-1a737209af70-kube-api-access-59sp6\") pod \"e481cf82-f798-4da5-8ffc-1a737209af70\" (UID: \"e481cf82-f798-4da5-8ffc-1a737209af70\") " Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.550163 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e481cf82-f798-4da5-8ffc-1a737209af70-logs" (OuterVolumeSpecName: "logs") pod "e481cf82-f798-4da5-8ffc-1a737209af70" (UID: "e481cf82-f798-4da5-8ffc-1a737209af70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.550821 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e481cf82-f798-4da5-8ffc-1a737209af70-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.555232 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e481cf82-f798-4da5-8ffc-1a737209af70-kube-api-access-59sp6" (OuterVolumeSpecName: "kube-api-access-59sp6") pod "e481cf82-f798-4da5-8ffc-1a737209af70" (UID: "e481cf82-f798-4da5-8ffc-1a737209af70"). InnerVolumeSpecName "kube-api-access-59sp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.576977 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e481cf82-f798-4da5-8ffc-1a737209af70" (UID: "e481cf82-f798-4da5-8ffc-1a737209af70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.591943 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-config-data" (OuterVolumeSpecName: "config-data") pod "e481cf82-f798-4da5-8ffc-1a737209af70" (UID: "e481cf82-f798-4da5-8ffc-1a737209af70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.613502 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e481cf82-f798-4da5-8ffc-1a737209af70" (UID: "e481cf82-f798-4da5-8ffc-1a737209af70"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.652891 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.653170 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.653239 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59sp6\" (UniqueName: \"kubernetes.io/projected/e481cf82-f798-4da5-8ffc-1a737209af70-kube-api-access-59sp6\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.653296 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e481cf82-f798-4da5-8ffc-1a737209af70-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.869885 4810 generic.go:334] "Generic (PLEG): container finished" podID="e481cf82-f798-4da5-8ffc-1a737209af70" containerID="e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e" exitCode=0 Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.869918 4810 generic.go:334] "Generic (PLEG): container finished" podID="e481cf82-f798-4da5-8ffc-1a737209af70" containerID="7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38" exitCode=143 Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.870036 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.871020 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e481cf82-f798-4da5-8ffc-1a737209af70","Type":"ContainerDied","Data":"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e"} Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.871052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e481cf82-f798-4da5-8ffc-1a737209af70","Type":"ContainerDied","Data":"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38"} Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.871063 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e481cf82-f798-4da5-8ffc-1a737209af70","Type":"ContainerDied","Data":"ac360ebccadf6d249bbcba2784c097de44f32a6b3efe1f43d0127688d29a2bbf"} Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.871078 4810 scope.go:117] "RemoveContainer" containerID="e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.874351 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b","Type":"ContainerStarted","Data":"f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e"} Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.875807 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.905090 4810 scope.go:117] "RemoveContainer" containerID="7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.909226 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.909194062 podStartE2EDuration="2.909194062s" podCreationTimestamp="2025-10-08 06:51:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:10.899033709 +0000 UTC m=+1173.533473449" watchObservedRunningTime="2025-10-08 06:51:10.909194062 +0000 UTC m=+1173.543633842" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.933906 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.947209 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.947269 4810 scope.go:117] "RemoveContainer" containerID="e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e" Oct 08 06:51:10 crc kubenswrapper[4810]: E1008 06:51:10.948165 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e\": container with ID starting with e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e not found: ID does not exist" containerID="e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.948214 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e"} err="failed to get container status \"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e\": rpc error: code = NotFound desc = could not find container \"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e\": container with ID starting with e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e not found: ID does not exist" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.948240 4810 scope.go:117] "RemoveContainer" containerID="7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38" Oct 08 06:51:10 crc kubenswrapper[4810]: E1008 06:51:10.949524 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38\": container with ID starting with 7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38 not found: ID does not exist" containerID="7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.949686 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38"} err="failed to get container status \"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38\": rpc error: code = NotFound desc = could not find container \"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38\": container with ID starting with 7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38 not found: ID does not exist" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.949825 4810 scope.go:117] "RemoveContainer" containerID="e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.952722 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e"} err="failed to get container status \"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e\": rpc error: code = NotFound desc = could not find container \"e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e\": container with ID starting with e55ef86e57a0d0a7aa4b44235dd28b44f9004cc6e09096bf7d3917c79160172e not found: ID does not exist" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.952769 4810 scope.go:117] "RemoveContainer" containerID="7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.953248 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38"} err="failed to get container status \"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38\": rpc error: code = NotFound desc = could not find container \"7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38\": container with ID starting with 7cda8e97ffb5a4a7f2fa7c30c6a7c08386db339157bc58ee8ddbe604aeb04a38 not found: ID does not exist" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.961386 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:10 crc kubenswrapper[4810]: E1008 06:51:10.961901 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-log" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.961922 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-log" Oct 08 06:51:10 crc kubenswrapper[4810]: E1008 06:51:10.961957 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-metadata" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.961984 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-metadata" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.962231 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-metadata" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.962266 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" containerName="nova-metadata-log" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.963567 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.967613 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.967668 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 06:51:10 crc kubenswrapper[4810]: I1008 06:51:10.972825 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.062232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.062305 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-config-data\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.062545 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244d0963-dd41-46da-89ad-8ea690f0f55d-logs\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.063031 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8mpw\" (UniqueName: \"kubernetes.io/projected/244d0963-dd41-46da-89ad-8ea690f0f55d-kube-api-access-z8mpw\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.063170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: E1008 06:51:11.131944 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 06:51:11 crc kubenswrapper[4810]: E1008 06:51:11.133670 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 06:51:11 crc kubenswrapper[4810]: E1008 06:51:11.134774 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 06:51:11 crc kubenswrapper[4810]: E1008 06:51:11.134807 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" containerName="nova-scheduler-scheduler" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.165231 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8mpw\" (UniqueName: \"kubernetes.io/projected/244d0963-dd41-46da-89ad-8ea690f0f55d-kube-api-access-z8mpw\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.165373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.165454 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.165496 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-config-data\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.165631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244d0963-dd41-46da-89ad-8ea690f0f55d-logs\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.168502 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244d0963-dd41-46da-89ad-8ea690f0f55d-logs\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.170641 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.170873 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.171048 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-config-data\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.182952 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8mpw\" (UniqueName: \"kubernetes.io/projected/244d0963-dd41-46da-89ad-8ea690f0f55d-kube-api-access-z8mpw\") pod \"nova-metadata-0\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.278815 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.756703 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:11 crc kubenswrapper[4810]: W1008 06:51:11.766658 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod244d0963_dd41_46da_89ad_8ea690f0f55d.slice/crio-d96b2c43cb06f0c7a16b0ffb962103590841f8a7bcccb70363190d8d2ba7c33b WatchSource:0}: Error finding container d96b2c43cb06f0c7a16b0ffb962103590841f8a7bcccb70363190d8d2ba7c33b: Status 404 returned error can't find the container with id d96b2c43cb06f0c7a16b0ffb962103590841f8a7bcccb70363190d8d2ba7c33b Oct 08 06:51:11 crc kubenswrapper[4810]: I1008 06:51:11.888294 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"244d0963-dd41-46da-89ad-8ea690f0f55d","Type":"ContainerStarted","Data":"d96b2c43cb06f0c7a16b0ffb962103590841f8a7bcccb70363190d8d2ba7c33b"} Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.086665 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e481cf82-f798-4da5-8ffc-1a737209af70" path="/var/lib/kubelet/pods/e481cf82-f798-4da5-8ffc-1a737209af70/volumes" Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.898886 4810 generic.go:334] "Generic (PLEG): container finished" podID="aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" containerID="83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0" exitCode=0 Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.899561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9","Type":"ContainerDied","Data":"83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0"} Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.901717 4810 generic.go:334] "Generic (PLEG): container finished" podID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerID="b1c8620d6969a132a8688a38415693d0c85acae6386ee9cee039b3d30043d082" exitCode=0 Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.901766 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf","Type":"ContainerDied","Data":"b1c8620d6969a132a8688a38415693d0c85acae6386ee9cee039b3d30043d082"} Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.901791 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf","Type":"ContainerDied","Data":"fbbac990da7329a33c1382cb9d1854733a890928a8001af02f1586417662c368"} Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.901802 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbbac990da7329a33c1382cb9d1854733a890928a8001af02f1586417662c368" Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.906499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"244d0963-dd41-46da-89ad-8ea690f0f55d","Type":"ContainerStarted","Data":"3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48"} Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.906537 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"244d0963-dd41-46da-89ad-8ea690f0f55d","Type":"ContainerStarted","Data":"3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633"} Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.942764 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9427444339999997 podStartE2EDuration="2.942744434s" podCreationTimestamp="2025-10-08 06:51:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:12.935885774 +0000 UTC m=+1175.570325524" watchObservedRunningTime="2025-10-08 06:51:12.942744434 +0000 UTC m=+1175.577184184" Oct 08 06:51:12 crc kubenswrapper[4810]: I1008 06:51:12.965209 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.005536 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-combined-ca-bundle\") pod \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.005684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgzr2\" (UniqueName: \"kubernetes.io/projected/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-kube-api-access-jgzr2\") pod \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.005718 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-config-data\") pod \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.005792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-logs\") pod \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\" (UID: \"e4b1bf70-6a1e-4c52-aa30-6b17c91397cf\") " Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.006601 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-logs" (OuterVolumeSpecName: "logs") pod "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" (UID: "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.007411 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.013992 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-kube-api-access-jgzr2" (OuterVolumeSpecName: "kube-api-access-jgzr2") pod "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" (UID: "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf"). InnerVolumeSpecName "kube-api-access-jgzr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.043935 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-config-data" (OuterVolumeSpecName: "config-data") pod "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" (UID: "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.104091 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" (UID: "e4b1bf70-6a1e-4c52-aa30-6b17c91397cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.111435 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.111486 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgzr2\" (UniqueName: \"kubernetes.io/projected/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-kube-api-access-jgzr2\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.111499 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.231087 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.314666 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-config-data\") pod \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.314711 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6w5z\" (UniqueName: \"kubernetes.io/projected/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-kube-api-access-w6w5z\") pod \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.314979 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-combined-ca-bundle\") pod \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\" (UID: \"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9\") " Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.318704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-kube-api-access-w6w5z" (OuterVolumeSpecName: "kube-api-access-w6w5z") pod "aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" (UID: "aba19f5a-3c12-4ec3-a567-4f326aaf2fc9"). InnerVolumeSpecName "kube-api-access-w6w5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.344038 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-config-data" (OuterVolumeSpecName: "config-data") pod "aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" (UID: "aba19f5a-3c12-4ec3-a567-4f326aaf2fc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.354278 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" (UID: "aba19f5a-3c12-4ec3-a567-4f326aaf2fc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.417783 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.417826 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6w5z\" (UniqueName: \"kubernetes.io/projected/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-kube-api-access-w6w5z\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.417841 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.916549 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aba19f5a-3c12-4ec3-a567-4f326aaf2fc9","Type":"ContainerDied","Data":"977432afab6d0cfd25f035fbb91479bfbb3b320b5d0e56f3aea105a1e9bba45a"} Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.917749 4810 scope.go:117] "RemoveContainer" containerID="83896c664bb90d7542ec8dd1ee711af6671d3de6a090974b50b661dca77092a0" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.916634 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.916623 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.961766 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:13 crc kubenswrapper[4810]: I1008 06:51:13.978342 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.001045 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.015085 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:14 crc kubenswrapper[4810]: E1008 06:51:14.015639 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" containerName="nova-scheduler-scheduler" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.015662 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" containerName="nova-scheduler-scheduler" Oct 08 06:51:14 crc kubenswrapper[4810]: E1008 06:51:14.015702 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-log" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.015710 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-log" Oct 08 06:51:14 crc kubenswrapper[4810]: E1008 06:51:14.015719 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-api" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.015725 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-api" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.015895 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-log" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.015921 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" containerName="nova-scheduler-scheduler" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.015929 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" containerName="nova-api-api" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.016721 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.019545 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.024353 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.033427 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.046070 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.048015 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.052078 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.056060 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.086159 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aba19f5a-3c12-4ec3-a567-4f326aaf2fc9" path="/var/lib/kubelet/pods/aba19f5a-3c12-4ec3-a567-4f326aaf2fc9/volumes" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.087014 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4b1bf70-6a1e-4c52-aa30-6b17c91397cf" path="/var/lib/kubelet/pods/e4b1bf70-6a1e-4c52-aa30-6b17c91397cf/volumes" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.130840 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c461df2a-a24d-4ec2-b964-161d0c263048-logs\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.131233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.131408 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8bhn\" (UniqueName: \"kubernetes.io/projected/36e4c82b-30d2-4e56-8b1a-9646115da8b2-kube-api-access-t8bhn\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.131555 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-config-data\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.131670 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.131793 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-config-data\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.131936 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d48q\" (UniqueName: \"kubernetes.io/projected/c461df2a-a24d-4ec2-b964-161d0c263048-kube-api-access-9d48q\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.234265 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d48q\" (UniqueName: \"kubernetes.io/projected/c461df2a-a24d-4ec2-b964-161d0c263048-kube-api-access-9d48q\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.234628 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c461df2a-a24d-4ec2-b964-161d0c263048-logs\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.234827 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.235154 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c461df2a-a24d-4ec2-b964-161d0c263048-logs\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.235189 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8bhn\" (UniqueName: \"kubernetes.io/projected/36e4c82b-30d2-4e56-8b1a-9646115da8b2-kube-api-access-t8bhn\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.235297 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-config-data\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.235325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.235378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-config-data\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.248516 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.249109 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.249680 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-config-data\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.251986 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8bhn\" (UniqueName: \"kubernetes.io/projected/36e4c82b-30d2-4e56-8b1a-9646115da8b2-kube-api-access-t8bhn\") pod \"nova-scheduler-0\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.252736 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.254536 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-config-data\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.255408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d48q\" (UniqueName: \"kubernetes.io/projected/c461df2a-a24d-4ec2-b964-161d0c263048-kube-api-access-9d48q\") pod \"nova-api-0\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " pod="openstack/nova-api-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.334850 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:14 crc kubenswrapper[4810]: I1008 06:51:14.372808 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:14.874594 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:15 crc kubenswrapper[4810]: W1008 06:51:14.876666 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36e4c82b_30d2_4e56_8b1a_9646115da8b2.slice/crio-c45bfc2b71d71ca54b4364c9d436851915d7670e3a336cbe6f35e7b7e652c483 WatchSource:0}: Error finding container c45bfc2b71d71ca54b4364c9d436851915d7670e3a336cbe6f35e7b7e652c483: Status 404 returned error can't find the container with id c45bfc2b71d71ca54b4364c9d436851915d7670e3a336cbe6f35e7b7e652c483 Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:14.932704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36e4c82b-30d2-4e56-8b1a-9646115da8b2","Type":"ContainerStarted","Data":"c45bfc2b71d71ca54b4364c9d436851915d7670e3a336cbe6f35e7b7e652c483"} Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:15.003998 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:15.946322 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36e4c82b-30d2-4e56-8b1a-9646115da8b2","Type":"ContainerStarted","Data":"e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7"} Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:15.950334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c461df2a-a24d-4ec2-b964-161d0c263048","Type":"ContainerStarted","Data":"e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9"} Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:15.950381 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c461df2a-a24d-4ec2-b964-161d0c263048","Type":"ContainerStarted","Data":"714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759"} Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:15.950390 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c461df2a-a24d-4ec2-b964-161d0c263048","Type":"ContainerStarted","Data":"b1f1216d1ce363e67d9682d4e07d01c82987442e9d2b819e66652abdb3499765"} Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:15.976536 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.9765000280000002 podStartE2EDuration="2.976500028s" podCreationTimestamp="2025-10-08 06:51:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:15.965338858 +0000 UTC m=+1178.599778598" watchObservedRunningTime="2025-10-08 06:51:15.976500028 +0000 UTC m=+1178.610939768" Oct 08 06:51:15 crc kubenswrapper[4810]: I1008 06:51:15.992238 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.992220124 podStartE2EDuration="2.992220124s" podCreationTimestamp="2025-10-08 06:51:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:15.988777328 +0000 UTC m=+1178.623217098" watchObservedRunningTime="2025-10-08 06:51:15.992220124 +0000 UTC m=+1178.626659864" Oct 08 06:51:16 crc kubenswrapper[4810]: I1008 06:51:16.279911 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 06:51:16 crc kubenswrapper[4810]: I1008 06:51:16.280886 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 06:51:19 crc kubenswrapper[4810]: I1008 06:51:19.335754 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 06:51:19 crc kubenswrapper[4810]: I1008 06:51:19.814617 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 06:51:21 crc kubenswrapper[4810]: I1008 06:51:21.279987 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 06:51:21 crc kubenswrapper[4810]: I1008 06:51:21.280238 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 06:51:22 crc kubenswrapper[4810]: I1008 06:51:22.298071 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 06:51:22 crc kubenswrapper[4810]: I1008 06:51:22.298395 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 06:51:24 crc kubenswrapper[4810]: I1008 06:51:24.336075 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 06:51:24 crc kubenswrapper[4810]: I1008 06:51:24.374075 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 06:51:24 crc kubenswrapper[4810]: I1008 06:51:24.374132 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 06:51:24 crc kubenswrapper[4810]: I1008 06:51:24.377214 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 06:51:25 crc kubenswrapper[4810]: I1008 06:51:25.062904 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 06:51:25 crc kubenswrapper[4810]: I1008 06:51:25.456224 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 06:51:25 crc kubenswrapper[4810]: I1008 06:51:25.456224 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 06:51:31 crc kubenswrapper[4810]: I1008 06:51:31.286181 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 06:51:31 crc kubenswrapper[4810]: I1008 06:51:31.286806 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 06:51:31 crc kubenswrapper[4810]: I1008 06:51:31.295652 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 06:51:31 crc kubenswrapper[4810]: I1008 06:51:31.298158 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.034275 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.103639 4810 generic.go:334] "Generic (PLEG): container finished" podID="1adfe378-7e5d-4274-a5c5-0ab624c84bcf" containerID="2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763" exitCode=137 Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.103697 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1adfe378-7e5d-4274-a5c5-0ab624c84bcf","Type":"ContainerDied","Data":"2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763"} Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.103751 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1adfe378-7e5d-4274-a5c5-0ab624c84bcf","Type":"ContainerDied","Data":"49c4adbec3c13d78a266ea6df1dab9a03eec80e30eb0b8f12f7d2fd46e9dec27"} Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.103772 4810 scope.go:117] "RemoveContainer" containerID="2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.103712 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.133291 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-combined-ca-bundle\") pod \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.133496 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txfdl\" (UniqueName: \"kubernetes.io/projected/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-kube-api-access-txfdl\") pod \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.133542 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-config-data\") pod \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\" (UID: \"1adfe378-7e5d-4274-a5c5-0ab624c84bcf\") " Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.133924 4810 scope.go:117] "RemoveContainer" containerID="2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763" Oct 08 06:51:32 crc kubenswrapper[4810]: E1008 06:51:32.134388 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763\": container with ID starting with 2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763 not found: ID does not exist" containerID="2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.134435 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763"} err="failed to get container status \"2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763\": rpc error: code = NotFound desc = could not find container \"2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763\": container with ID starting with 2175270fe79f68dd550be4f7c2e92f6d79d07290738fbde97770526f704c5763 not found: ID does not exist" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.141338 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-kube-api-access-txfdl" (OuterVolumeSpecName: "kube-api-access-txfdl") pod "1adfe378-7e5d-4274-a5c5-0ab624c84bcf" (UID: "1adfe378-7e5d-4274-a5c5-0ab624c84bcf"). InnerVolumeSpecName "kube-api-access-txfdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.160106 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-config-data" (OuterVolumeSpecName: "config-data") pod "1adfe378-7e5d-4274-a5c5-0ab624c84bcf" (UID: "1adfe378-7e5d-4274-a5c5-0ab624c84bcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.162563 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1adfe378-7e5d-4274-a5c5-0ab624c84bcf" (UID: "1adfe378-7e5d-4274-a5c5-0ab624c84bcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.238192 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txfdl\" (UniqueName: \"kubernetes.io/projected/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-kube-api-access-txfdl\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.238226 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.238238 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1adfe378-7e5d-4274-a5c5-0ab624c84bcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.433584 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.442253 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.455943 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:51:32 crc kubenswrapper[4810]: E1008 06:51:32.456528 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1adfe378-7e5d-4274-a5c5-0ab624c84bcf" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.456552 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1adfe378-7e5d-4274-a5c5-0ab624c84bcf" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.456873 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1adfe378-7e5d-4274-a5c5-0ab624c84bcf" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.457789 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.461197 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.461448 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.461565 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.471696 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.543577 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5t59\" (UniqueName: \"kubernetes.io/projected/9895842c-0b86-4c8c-9076-3777d45634dd-kube-api-access-r5t59\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.543700 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.543731 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.543764 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.543783 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.645451 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5t59\" (UniqueName: \"kubernetes.io/projected/9895842c-0b86-4c8c-9076-3777d45634dd-kube-api-access-r5t59\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.645575 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.645602 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.645632 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.645652 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.649635 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.649792 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.650008 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.650095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.662215 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5t59\" (UniqueName: \"kubernetes.io/projected/9895842c-0b86-4c8c-9076-3777d45634dd-kube-api-access-r5t59\") pod \"nova-cell1-novncproxy-0\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:32 crc kubenswrapper[4810]: I1008 06:51:32.788512 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:33 crc kubenswrapper[4810]: I1008 06:51:33.269679 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:51:33 crc kubenswrapper[4810]: W1008 06:51:33.272131 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9895842c_0b86_4c8c_9076_3777d45634dd.slice/crio-74aec784c769a9420a25956a07e2b3353e806c230beaed1a19347a590ebb6112 WatchSource:0}: Error finding container 74aec784c769a9420a25956a07e2b3353e806c230beaed1a19347a590ebb6112: Status 404 returned error can't find the container with id 74aec784c769a9420a25956a07e2b3353e806c230beaed1a19347a590ebb6112 Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.087519 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1adfe378-7e5d-4274-a5c5-0ab624c84bcf" path="/var/lib/kubelet/pods/1adfe378-7e5d-4274-a5c5-0ab624c84bcf/volumes" Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.137249 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9895842c-0b86-4c8c-9076-3777d45634dd","Type":"ContainerStarted","Data":"f995d871a3f41fdb1cb9c3fea81de2923a198870ea13f734a17d0294e8d0433f"} Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.137321 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9895842c-0b86-4c8c-9076-3777d45634dd","Type":"ContainerStarted","Data":"74aec784c769a9420a25956a07e2b3353e806c230beaed1a19347a590ebb6112"} Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.176151 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.17611972 podStartE2EDuration="2.17611972s" podCreationTimestamp="2025-10-08 06:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:34.159260752 +0000 UTC m=+1196.793700492" watchObservedRunningTime="2025-10-08 06:51:34.17611972 +0000 UTC m=+1196.810559500" Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.379244 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.379821 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.382540 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 06:51:34 crc kubenswrapper[4810]: I1008 06:51:34.383221 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.144696 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.149975 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.357292 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d4d96bb9-89r5m"] Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.360282 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.382933 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d4d96bb9-89r5m"] Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.408233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-sb\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.408605 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-nb\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.408782 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-swift-storage-0\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.408873 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-svc\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.409025 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-config\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.409188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm77d\" (UniqueName: \"kubernetes.io/projected/5091c146-f092-4878-95a3-49cb91760d17-kube-api-access-dm77d\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.511297 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-config\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.511393 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm77d\" (UniqueName: \"kubernetes.io/projected/5091c146-f092-4878-95a3-49cb91760d17-kube-api-access-dm77d\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.511464 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-sb\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.511494 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-nb\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.511556 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-svc\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.511571 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-swift-storage-0\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.512535 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-config\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.512666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-sb\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.512798 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-nb\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.512866 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-svc\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.513212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-swift-storage-0\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.532717 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm77d\" (UniqueName: \"kubernetes.io/projected/5091c146-f092-4878-95a3-49cb91760d17-kube-api-access-dm77d\") pod \"dnsmasq-dns-6d4d96bb9-89r5m\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:35 crc kubenswrapper[4810]: I1008 06:51:35.681619 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:36 crc kubenswrapper[4810]: I1008 06:51:36.175981 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d4d96bb9-89r5m"] Oct 08 06:51:36 crc kubenswrapper[4810]: W1008 06:51:36.182755 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5091c146_f092_4878_95a3_49cb91760d17.slice/crio-4c5faeb23f4543d30881f9452b47e01d9f2e9be61c6a1a433d8bd4b1d0f9d199 WatchSource:0}: Error finding container 4c5faeb23f4543d30881f9452b47e01d9f2e9be61c6a1a433d8bd4b1d0f9d199: Status 404 returned error can't find the container with id 4c5faeb23f4543d30881f9452b47e01d9f2e9be61c6a1a433d8bd4b1d0f9d199 Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.165558 4810 generic.go:334] "Generic (PLEG): container finished" podID="5091c146-f092-4878-95a3-49cb91760d17" containerID="2ae3f130783d76beb259a6e72a55a197e29b739bcfbe13913f1f50be2fef28b4" exitCode=0 Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.165664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" event={"ID":"5091c146-f092-4878-95a3-49cb91760d17","Type":"ContainerDied","Data":"2ae3f130783d76beb259a6e72a55a197e29b739bcfbe13913f1f50be2fef28b4"} Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.166153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" event={"ID":"5091c146-f092-4878-95a3-49cb91760d17","Type":"ContainerStarted","Data":"4c5faeb23f4543d30881f9452b47e01d9f2e9be61c6a1a433d8bd4b1d0f9d199"} Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.314404 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.314752 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-central-agent" containerID="cri-o://9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451" gracePeriod=30 Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.314898 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="proxy-httpd" containerID="cri-o://c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd" gracePeriod=30 Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.314944 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="sg-core" containerID="cri-o://569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec" gracePeriod=30 Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.315010 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-notification-agent" containerID="cri-o://54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b" gracePeriod=30 Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.589484 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:37 crc kubenswrapper[4810]: I1008 06:51:37.789223 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.181833 4810 generic.go:334] "Generic (PLEG): container finished" podID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerID="c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd" exitCode=0 Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.181878 4810 generic.go:334] "Generic (PLEG): container finished" podID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerID="569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec" exitCode=2 Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.181887 4810 generic.go:334] "Generic (PLEG): container finished" podID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerID="9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451" exitCode=0 Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.181914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerDied","Data":"c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd"} Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.181981 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerDied","Data":"569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec"} Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.182000 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerDied","Data":"9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451"} Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.184690 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-log" containerID="cri-o://714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759" gracePeriod=30 Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.185666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" event={"ID":"5091c146-f092-4878-95a3-49cb91760d17","Type":"ContainerStarted","Data":"7d7302d109e2280340776457b1d206311a58db6e8a68f517aea3acb2a3267097"} Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.185736 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-api" containerID="cri-o://e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9" gracePeriod=30 Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.186020 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:38 crc kubenswrapper[4810]: I1008 06:51:38.212485 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" podStartSLOduration=3.21245206 podStartE2EDuration="3.21245206s" podCreationTimestamp="2025-10-08 06:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:38.201904198 +0000 UTC m=+1200.836343938" watchObservedRunningTime="2025-10-08 06:51:38.21245206 +0000 UTC m=+1200.846891800" Oct 08 06:51:39 crc kubenswrapper[4810]: I1008 06:51:39.194531 4810 generic.go:334] "Generic (PLEG): container finished" podID="c461df2a-a24d-4ec2-b964-161d0c263048" containerID="714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759" exitCode=143 Oct 08 06:51:39 crc kubenswrapper[4810]: I1008 06:51:39.194635 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c461df2a-a24d-4ec2-b964-161d0c263048","Type":"ContainerDied","Data":"714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759"} Oct 08 06:51:41 crc kubenswrapper[4810]: I1008 06:51:41.965027 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.060441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-config-data\") pod \"c461df2a-a24d-4ec2-b964-161d0c263048\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.060571 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d48q\" (UniqueName: \"kubernetes.io/projected/c461df2a-a24d-4ec2-b964-161d0c263048-kube-api-access-9d48q\") pod \"c461df2a-a24d-4ec2-b964-161d0c263048\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.060687 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c461df2a-a24d-4ec2-b964-161d0c263048-logs\") pod \"c461df2a-a24d-4ec2-b964-161d0c263048\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.060829 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-combined-ca-bundle\") pod \"c461df2a-a24d-4ec2-b964-161d0c263048\" (UID: \"c461df2a-a24d-4ec2-b964-161d0c263048\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.061657 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c461df2a-a24d-4ec2-b964-161d0c263048-logs" (OuterVolumeSpecName: "logs") pod "c461df2a-a24d-4ec2-b964-161d0c263048" (UID: "c461df2a-a24d-4ec2-b964-161d0c263048"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.069448 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c461df2a-a24d-4ec2-b964-161d0c263048-kube-api-access-9d48q" (OuterVolumeSpecName: "kube-api-access-9d48q") pod "c461df2a-a24d-4ec2-b964-161d0c263048" (UID: "c461df2a-a24d-4ec2-b964-161d0c263048"). InnerVolumeSpecName "kube-api-access-9d48q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.092231 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c461df2a-a24d-4ec2-b964-161d0c263048" (UID: "c461df2a-a24d-4ec2-b964-161d0c263048"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.096615 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.133220 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-config-data" (OuterVolumeSpecName: "config-data") pod "c461df2a-a24d-4ec2-b964-161d0c263048" (UID: "c461df2a-a24d-4ec2-b964-161d0c263048"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.162426 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-sg-core-conf-yaml\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.162471 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-ceilometer-tls-certs\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.162594 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-combined-ca-bundle\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.162776 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-config-data\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.162822 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzhcg\" (UniqueName: \"kubernetes.io/projected/c8854e99-28ee-4a93-b755-f362a8bf1c51-kube-api-access-kzhcg\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.162871 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-run-httpd\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.162956 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-scripts\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.163006 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-log-httpd\") pod \"c8854e99-28ee-4a93-b755-f362a8bf1c51\" (UID: \"c8854e99-28ee-4a93-b755-f362a8bf1c51\") " Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.163575 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.163591 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d48q\" (UniqueName: \"kubernetes.io/projected/c461df2a-a24d-4ec2-b964-161d0c263048-kube-api-access-9d48q\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.163604 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c461df2a-a24d-4ec2-b964-161d0c263048-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.163615 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c461df2a-a24d-4ec2-b964-161d0c263048-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.164386 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.164615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.170216 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8854e99-28ee-4a93-b755-f362a8bf1c51-kube-api-access-kzhcg" (OuterVolumeSpecName: "kube-api-access-kzhcg") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "kube-api-access-kzhcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.170378 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-scripts" (OuterVolumeSpecName: "scripts") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.196959 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.229015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.232677 4810 generic.go:334] "Generic (PLEG): container finished" podID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerID="54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b" exitCode=0 Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.232742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerDied","Data":"54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b"} Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.232772 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8854e99-28ee-4a93-b755-f362a8bf1c51","Type":"ContainerDied","Data":"17d7a0799428358e0244666f47bb8077ae4eddc345c36b31374852b3df9f3ff0"} Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.232791 4810 scope.go:117] "RemoveContainer" containerID="c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.232925 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.240440 4810 generic.go:334] "Generic (PLEG): container finished" podID="c461df2a-a24d-4ec2-b964-161d0c263048" containerID="e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9" exitCode=0 Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.240484 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c461df2a-a24d-4ec2-b964-161d0c263048","Type":"ContainerDied","Data":"e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9"} Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.240514 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c461df2a-a24d-4ec2-b964-161d0c263048","Type":"ContainerDied","Data":"b1f1216d1ce363e67d9682d4e07d01c82987442e9d2b819e66652abdb3499765"} Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.240642 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.265292 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzhcg\" (UniqueName: \"kubernetes.io/projected/c8854e99-28ee-4a93-b755-f362a8bf1c51-kube-api-access-kzhcg\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.265347 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.265357 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.265366 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8854e99-28ee-4a93-b755-f362a8bf1c51-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.265402 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.265414 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.281501 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.286522 4810 scope.go:117] "RemoveContainer" containerID="569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.291338 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.307097 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.324780 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.325306 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-notification-agent" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325320 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-notification-agent" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.325343 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="sg-core" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325351 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="sg-core" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.325370 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-api" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325376 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-api" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.325399 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-log" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325405 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-log" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.325416 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="proxy-httpd" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325422 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="proxy-httpd" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.325433 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-central-agent" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325440 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-central-agent" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325657 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-central-agent" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325731 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="sg-core" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325756 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-api" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325766 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" containerName="nova-api-log" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325783 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="ceilometer-notification-agent" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.325798 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" containerName="proxy-httpd" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.327019 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.329446 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.330422 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.331867 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.333798 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.339503 4810 scope.go:117] "RemoveContainer" containerID="54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.340101 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-config-data" (OuterVolumeSpecName: "config-data") pod "c8854e99-28ee-4a93-b755-f362a8bf1c51" (UID: "c8854e99-28ee-4a93-b755-f362a8bf1c51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367240 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-config-data\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367302 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367351 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t58vx\" (UniqueName: \"kubernetes.io/projected/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-kube-api-access-t58vx\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367384 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-logs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367466 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-public-tls-certs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367610 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.367621 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8854e99-28ee-4a93-b755-f362a8bf1c51-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.378849 4810 scope.go:117] "RemoveContainer" containerID="9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.407089 4810 scope.go:117] "RemoveContainer" containerID="c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.411528 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd\": container with ID starting with c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd not found: ID does not exist" containerID="c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.411593 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd"} err="failed to get container status \"c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd\": rpc error: code = NotFound desc = could not find container \"c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd\": container with ID starting with c9abb9b41861ca3e3b038746971a9da6deeb6cf8f3fddc1e1d4f1b621963effd not found: ID does not exist" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.411628 4810 scope.go:117] "RemoveContainer" containerID="569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.412470 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec\": container with ID starting with 569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec not found: ID does not exist" containerID="569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.412526 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec"} err="failed to get container status \"569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec\": rpc error: code = NotFound desc = could not find container \"569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec\": container with ID starting with 569b3e8782ab755b706ad17aa8f162ff4277f2cfde28c0624e375adee40d99ec not found: ID does not exist" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.412560 4810 scope.go:117] "RemoveContainer" containerID="54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.417328 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b\": container with ID starting with 54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b not found: ID does not exist" containerID="54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.417380 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b"} err="failed to get container status \"54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b\": rpc error: code = NotFound desc = could not find container \"54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b\": container with ID starting with 54ece0ac4ef9fd6f98c96ab6c834ce32a29098ed918fadfbd338ce03c14fe45b not found: ID does not exist" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.417412 4810 scope.go:117] "RemoveContainer" containerID="9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.417776 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451\": container with ID starting with 9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451 not found: ID does not exist" containerID="9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.417805 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451"} err="failed to get container status \"9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451\": rpc error: code = NotFound desc = could not find container \"9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451\": container with ID starting with 9bbfe23658704ab57df777aa0da736b997cccb40410b10b5553e15875a4d1451 not found: ID does not exist" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.417822 4810 scope.go:117] "RemoveContainer" containerID="e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.469816 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-config-data\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.470334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.470374 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t58vx\" (UniqueName: \"kubernetes.io/projected/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-kube-api-access-t58vx\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.470415 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-logs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.470509 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.470561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-public-tls-certs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.472147 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-logs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.474356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-public-tls-certs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.475564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-config-data\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.476555 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.485442 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.486784 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t58vx\" (UniqueName: \"kubernetes.io/projected/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-kube-api-access-t58vx\") pod \"nova-api-0\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.492512 4810 scope.go:117] "RemoveContainer" containerID="714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.570702 4810 scope.go:117] "RemoveContainer" containerID="e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.572010 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9\": container with ID starting with e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9 not found: ID does not exist" containerID="e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.572048 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9"} err="failed to get container status \"e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9\": rpc error: code = NotFound desc = could not find container \"e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9\": container with ID starting with e3e02a048af3602c001394c0f84a04cbbe5791ace497a1474e9d838a6cd7fdc9 not found: ID does not exist" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.572073 4810 scope.go:117] "RemoveContainer" containerID="714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759" Oct 08 06:51:42 crc kubenswrapper[4810]: E1008 06:51:42.577243 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759\": container with ID starting with 714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759 not found: ID does not exist" containerID="714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.577299 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759"} err="failed to get container status \"714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759\": rpc error: code = NotFound desc = could not find container \"714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759\": container with ID starting with 714f4428c6f9aeda050cba935b190504b4666b6f05a1c6f11e929bcfe8e4c759 not found: ID does not exist" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.578669 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.590901 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.602207 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.604728 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.611568 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.612069 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.612259 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.619498 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.658053 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.673399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-run-httpd\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.673653 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-config-data\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.673762 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-log-httpd\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.673941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wqcf\" (UniqueName: \"kubernetes.io/projected/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-kube-api-access-8wqcf\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.674052 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.674180 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.674282 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.674400 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-scripts\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-run-httpd\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776539 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-config-data\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776572 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-log-httpd\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776630 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wqcf\" (UniqueName: \"kubernetes.io/projected/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-kube-api-access-8wqcf\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776651 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776671 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776701 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.776737 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-scripts\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.777269 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-run-httpd\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.777595 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-log-httpd\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.783193 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-scripts\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.783456 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-config-data\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.783266 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.786236 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.786568 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.789034 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.799136 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wqcf\" (UniqueName: \"kubernetes.io/projected/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-kube-api-access-8wqcf\") pod \"ceilometer-0\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " pod="openstack/ceilometer-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.822455 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:42 crc kubenswrapper[4810]: I1008 06:51:42.931548 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.118555 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.252521 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"052c9e6d-bec3-4249-8c2a-742b4eea9aa1","Type":"ContainerStarted","Data":"b503fa6b34fa3cdf07321eb7c3d42671ffab82eaee6a0c7bf86e57960bca8125"} Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.270392 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.401822 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.517537 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-vfmcm"] Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.524351 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.530466 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.536053 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.551433 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vfmcm"] Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.603060 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-scripts\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.603131 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-config-data\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.603185 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wzbc\" (UniqueName: \"kubernetes.io/projected/a5f82549-5e55-491b-9c01-bca9f29e25f3-kube-api-access-7wzbc\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.603251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.705446 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-scripts\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.705535 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-config-data\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.705590 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wzbc\" (UniqueName: \"kubernetes.io/projected/a5f82549-5e55-491b-9c01-bca9f29e25f3-kube-api-access-7wzbc\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.705656 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.715757 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.727983 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wzbc\" (UniqueName: \"kubernetes.io/projected/a5f82549-5e55-491b-9c01-bca9f29e25f3-kube-api-access-7wzbc\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.732711 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-scripts\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.733559 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-config-data\") pod \"nova-cell1-cell-mapping-vfmcm\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:43 crc kubenswrapper[4810]: I1008 06:51:43.875552 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:44 crc kubenswrapper[4810]: I1008 06:51:44.088632 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c461df2a-a24d-4ec2-b964-161d0c263048" path="/var/lib/kubelet/pods/c461df2a-a24d-4ec2-b964-161d0c263048/volumes" Oct 08 06:51:44 crc kubenswrapper[4810]: I1008 06:51:44.089998 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8854e99-28ee-4a93-b755-f362a8bf1c51" path="/var/lib/kubelet/pods/c8854e99-28ee-4a93-b755-f362a8bf1c51/volumes" Oct 08 06:51:44 crc kubenswrapper[4810]: I1008 06:51:44.264242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerStarted","Data":"96aa368956d59be4b085c7e5b85434c43547f5720ca062f547c74939c7f33876"} Oct 08 06:51:44 crc kubenswrapper[4810]: I1008 06:51:44.269505 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"052c9e6d-bec3-4249-8c2a-742b4eea9aa1","Type":"ContainerStarted","Data":"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc"} Oct 08 06:51:44 crc kubenswrapper[4810]: I1008 06:51:44.269577 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"052c9e6d-bec3-4249-8c2a-742b4eea9aa1","Type":"ContainerStarted","Data":"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8"} Oct 08 06:51:44 crc kubenswrapper[4810]: I1008 06:51:44.303941 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.303914922 podStartE2EDuration="2.303914922s" podCreationTimestamp="2025-10-08 06:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:44.286796447 +0000 UTC m=+1206.921236197" watchObservedRunningTime="2025-10-08 06:51:44.303914922 +0000 UTC m=+1206.938354662" Oct 08 06:51:44 crc kubenswrapper[4810]: I1008 06:51:44.447240 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vfmcm"] Oct 08 06:51:44 crc kubenswrapper[4810]: W1008 06:51:44.467364 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5f82549_5e55_491b_9c01_bca9f29e25f3.slice/crio-f61a61e869c71807aa21305f0e8f197473af6be363b9172f8908119708a7e0fe WatchSource:0}: Error finding container f61a61e869c71807aa21305f0e8f197473af6be363b9172f8908119708a7e0fe: Status 404 returned error can't find the container with id f61a61e869c71807aa21305f0e8f197473af6be363b9172f8908119708a7e0fe Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.279553 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vfmcm" event={"ID":"a5f82549-5e55-491b-9c01-bca9f29e25f3","Type":"ContainerStarted","Data":"2a383918c50471b7a1bb4a62090441f5f687cf37ba900114f603b1c349f14dcb"} Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.280110 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vfmcm" event={"ID":"a5f82549-5e55-491b-9c01-bca9f29e25f3","Type":"ContainerStarted","Data":"f61a61e869c71807aa21305f0e8f197473af6be363b9172f8908119708a7e0fe"} Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.282993 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerStarted","Data":"6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c"} Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.283044 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerStarted","Data":"1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495"} Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.300479 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-vfmcm" podStartSLOduration=2.3004576119999998 podStartE2EDuration="2.300457612s" podCreationTimestamp="2025-10-08 06:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:45.291219285 +0000 UTC m=+1207.925659025" watchObservedRunningTime="2025-10-08 06:51:45.300457612 +0000 UTC m=+1207.934897352" Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.684277 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.785521 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffc974fdf-6dtwt"] Oct 08 06:51:45 crc kubenswrapper[4810]: I1008 06:51:45.785880 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerName="dnsmasq-dns" containerID="cri-o://94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339" gracePeriod=10 Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.294720 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerStarted","Data":"9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049"} Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.295140 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.296280 4810 generic.go:334] "Generic (PLEG): container finished" podID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerID="94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339" exitCode=0 Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.296378 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" event={"ID":"89c96d63-b16e-40bb-9acd-a60a85cd8af2","Type":"ContainerDied","Data":"94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339"} Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.296430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" event={"ID":"89c96d63-b16e-40bb-9acd-a60a85cd8af2","Type":"ContainerDied","Data":"b8e47bdb3ef319497fa8bc576caf7b6dddbf1acae2053881413906fe60bb8914"} Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.296451 4810 scope.go:117] "RemoveContainer" containerID="94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.329738 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-svc\") pod \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.329785 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzvlv\" (UniqueName: \"kubernetes.io/projected/89c96d63-b16e-40bb-9acd-a60a85cd8af2-kube-api-access-bzvlv\") pod \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.329836 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-sb\") pod \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.329885 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-config\") pod \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.329914 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-nb\") pod \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.329982 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-swift-storage-0\") pod \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\" (UID: \"89c96d63-b16e-40bb-9acd-a60a85cd8af2\") " Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.360419 4810 scope.go:117] "RemoveContainer" containerID="fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.364697 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89c96d63-b16e-40bb-9acd-a60a85cd8af2-kube-api-access-bzvlv" (OuterVolumeSpecName: "kube-api-access-bzvlv") pod "89c96d63-b16e-40bb-9acd-a60a85cd8af2" (UID: "89c96d63-b16e-40bb-9acd-a60a85cd8af2"). InnerVolumeSpecName "kube-api-access-bzvlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.405218 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "89c96d63-b16e-40bb-9acd-a60a85cd8af2" (UID: "89c96d63-b16e-40bb-9acd-a60a85cd8af2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.412556 4810 scope.go:117] "RemoveContainer" containerID="94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339" Oct 08 06:51:46 crc kubenswrapper[4810]: E1008 06:51:46.413450 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339\": container with ID starting with 94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339 not found: ID does not exist" containerID="94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.413484 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339"} err="failed to get container status \"94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339\": rpc error: code = NotFound desc = could not find container \"94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339\": container with ID starting with 94983cba383f76ff20c37afb69ff3287c469885956df255134b1e10cc881c339 not found: ID does not exist" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.413511 4810 scope.go:117] "RemoveContainer" containerID="fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751" Oct 08 06:51:46 crc kubenswrapper[4810]: E1008 06:51:46.417386 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751\": container with ID starting with fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751 not found: ID does not exist" containerID="fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.417427 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751"} err="failed to get container status \"fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751\": rpc error: code = NotFound desc = could not find container \"fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751\": container with ID starting with fbcef7d7ff36a7798a5012dced3ac82786b964c0cbe064d43f708269fafb0751 not found: ID does not exist" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.422633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "89c96d63-b16e-40bb-9acd-a60a85cd8af2" (UID: "89c96d63-b16e-40bb-9acd-a60a85cd8af2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.428384 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89c96d63-b16e-40bb-9acd-a60a85cd8af2" (UID: "89c96d63-b16e-40bb-9acd-a60a85cd8af2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.428869 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "89c96d63-b16e-40bb-9acd-a60a85cd8af2" (UID: "89c96d63-b16e-40bb-9acd-a60a85cd8af2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.432314 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.432410 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzvlv\" (UniqueName: \"kubernetes.io/projected/89c96d63-b16e-40bb-9acd-a60a85cd8af2-kube-api-access-bzvlv\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.432433 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.432446 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.432457 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.485597 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-config" (OuterVolumeSpecName: "config") pod "89c96d63-b16e-40bb-9acd-a60a85cd8af2" (UID: "89c96d63-b16e-40bb-9acd-a60a85cd8af2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:51:46 crc kubenswrapper[4810]: I1008 06:51:46.534182 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89c96d63-b16e-40bb-9acd-a60a85cd8af2-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:47 crc kubenswrapper[4810]: I1008 06:51:47.311848 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" Oct 08 06:51:47 crc kubenswrapper[4810]: I1008 06:51:47.317065 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerStarted","Data":"9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc"} Oct 08 06:51:47 crc kubenswrapper[4810]: I1008 06:51:47.318148 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 06:51:47 crc kubenswrapper[4810]: I1008 06:51:47.348662 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.761512073 podStartE2EDuration="5.34863983s" podCreationTimestamp="2025-10-08 06:51:42 +0000 UTC" firstStartedPulling="2025-10-08 06:51:43.432264538 +0000 UTC m=+1206.066704278" lastFinishedPulling="2025-10-08 06:51:47.019392305 +0000 UTC m=+1209.653832035" observedRunningTime="2025-10-08 06:51:47.339085924 +0000 UTC m=+1209.973525665" watchObservedRunningTime="2025-10-08 06:51:47.34863983 +0000 UTC m=+1209.983079570" Oct 08 06:51:47 crc kubenswrapper[4810]: I1008 06:51:47.366023 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffc974fdf-6dtwt"] Oct 08 06:51:47 crc kubenswrapper[4810]: I1008 06:51:47.373037 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6ffc974fdf-6dtwt"] Oct 08 06:51:48 crc kubenswrapper[4810]: I1008 06:51:48.083831 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" path="/var/lib/kubelet/pods/89c96d63-b16e-40bb-9acd-a60a85cd8af2/volumes" Oct 08 06:51:50 crc kubenswrapper[4810]: I1008 06:51:50.350385 4810 generic.go:334] "Generic (PLEG): container finished" podID="a5f82549-5e55-491b-9c01-bca9f29e25f3" containerID="2a383918c50471b7a1bb4a62090441f5f687cf37ba900114f603b1c349f14dcb" exitCode=0 Oct 08 06:51:50 crc kubenswrapper[4810]: I1008 06:51:50.350634 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vfmcm" event={"ID":"a5f82549-5e55-491b-9c01-bca9f29e25f3","Type":"ContainerDied","Data":"2a383918c50471b7a1bb4a62090441f5f687cf37ba900114f603b1c349f14dcb"} Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.207376 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6ffc974fdf-6dtwt" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.191:5353: i/o timeout" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.756727 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.849872 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-scripts\") pod \"a5f82549-5e55-491b-9c01-bca9f29e25f3\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.850177 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-config-data\") pod \"a5f82549-5e55-491b-9c01-bca9f29e25f3\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.850242 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wzbc\" (UniqueName: \"kubernetes.io/projected/a5f82549-5e55-491b-9c01-bca9f29e25f3-kube-api-access-7wzbc\") pod \"a5f82549-5e55-491b-9c01-bca9f29e25f3\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.850337 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-combined-ca-bundle\") pod \"a5f82549-5e55-491b-9c01-bca9f29e25f3\" (UID: \"a5f82549-5e55-491b-9c01-bca9f29e25f3\") " Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.866150 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f82549-5e55-491b-9c01-bca9f29e25f3-kube-api-access-7wzbc" (OuterVolumeSpecName: "kube-api-access-7wzbc") pod "a5f82549-5e55-491b-9c01-bca9f29e25f3" (UID: "a5f82549-5e55-491b-9c01-bca9f29e25f3"). InnerVolumeSpecName "kube-api-access-7wzbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.866637 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-scripts" (OuterVolumeSpecName: "scripts") pod "a5f82549-5e55-491b-9c01-bca9f29e25f3" (UID: "a5f82549-5e55-491b-9c01-bca9f29e25f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.891816 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-config-data" (OuterVolumeSpecName: "config-data") pod "a5f82549-5e55-491b-9c01-bca9f29e25f3" (UID: "a5f82549-5e55-491b-9c01-bca9f29e25f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.898242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5f82549-5e55-491b-9c01-bca9f29e25f3" (UID: "a5f82549-5e55-491b-9c01-bca9f29e25f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.953343 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.953393 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.953406 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f82549-5e55-491b-9c01-bca9f29e25f3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:51 crc kubenswrapper[4810]: I1008 06:51:51.953418 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wzbc\" (UniqueName: \"kubernetes.io/projected/a5f82549-5e55-491b-9c01-bca9f29e25f3-kube-api-access-7wzbc\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.378924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vfmcm" event={"ID":"a5f82549-5e55-491b-9c01-bca9f29e25f3","Type":"ContainerDied","Data":"f61a61e869c71807aa21305f0e8f197473af6be363b9172f8908119708a7e0fe"} Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.379012 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f61a61e869c71807aa21305f0e8f197473af6be363b9172f8908119708a7e0fe" Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.379074 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vfmcm" Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.616667 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.616945 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="36e4c82b-30d2-4e56-8b1a-9646115da8b2" containerName="nova-scheduler-scheduler" containerID="cri-o://e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7" gracePeriod=30 Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.636443 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.636785 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-log" containerID="cri-o://a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8" gracePeriod=30 Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.637017 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-api" containerID="cri-o://3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc" gracePeriod=30 Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.651493 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.651860 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-log" containerID="cri-o://3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633" gracePeriod=30 Oct 08 06:51:52 crc kubenswrapper[4810]: I1008 06:51:52.652040 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-metadata" containerID="cri-o://3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48" gracePeriod=30 Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.378326 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.390704 4810 generic.go:334] "Generic (PLEG): container finished" podID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerID="3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc" exitCode=0 Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.390734 4810 generic.go:334] "Generic (PLEG): container finished" podID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerID="a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8" exitCode=143 Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.390791 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"052c9e6d-bec3-4249-8c2a-742b4eea9aa1","Type":"ContainerDied","Data":"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc"} Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.390826 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"052c9e6d-bec3-4249-8c2a-742b4eea9aa1","Type":"ContainerDied","Data":"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8"} Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.390837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"052c9e6d-bec3-4249-8c2a-742b4eea9aa1","Type":"ContainerDied","Data":"b503fa6b34fa3cdf07321eb7c3d42671ffab82eaee6a0c7bf86e57960bca8125"} Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.390853 4810 scope.go:117] "RemoveContainer" containerID="3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.390851 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.392944 4810 generic.go:334] "Generic (PLEG): container finished" podID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerID="3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633" exitCode=143 Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.393011 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"244d0963-dd41-46da-89ad-8ea690f0f55d","Type":"ContainerDied","Data":"3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633"} Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.425774 4810 scope.go:117] "RemoveContainer" containerID="a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.499210 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-logs\") pod \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.499309 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-config-data\") pod \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.499374 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-internal-tls-certs\") pod \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.499630 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-public-tls-certs\") pod \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.499684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t58vx\" (UniqueName: \"kubernetes.io/projected/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-kube-api-access-t58vx\") pod \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.499701 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-combined-ca-bundle\") pod \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\" (UID: \"052c9e6d-bec3-4249-8c2a-742b4eea9aa1\") " Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.499995 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-logs" (OuterVolumeSpecName: "logs") pod "052c9e6d-bec3-4249-8c2a-742b4eea9aa1" (UID: "052c9e6d-bec3-4249-8c2a-742b4eea9aa1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.500421 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.516858 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-kube-api-access-t58vx" (OuterVolumeSpecName: "kube-api-access-t58vx") pod "052c9e6d-bec3-4249-8c2a-742b4eea9aa1" (UID: "052c9e6d-bec3-4249-8c2a-742b4eea9aa1"). InnerVolumeSpecName "kube-api-access-t58vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.524952 4810 scope.go:117] "RemoveContainer" containerID="3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.533616 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-config-data" (OuterVolumeSpecName: "config-data") pod "052c9e6d-bec3-4249-8c2a-742b4eea9aa1" (UID: "052c9e6d-bec3-4249-8c2a-742b4eea9aa1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:53 crc kubenswrapper[4810]: E1008 06:51:53.535803 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc\": container with ID starting with 3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc not found: ID does not exist" containerID="3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.535860 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc"} err="failed to get container status \"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc\": rpc error: code = NotFound desc = could not find container \"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc\": container with ID starting with 3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc not found: ID does not exist" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.535911 4810 scope.go:117] "RemoveContainer" containerID="a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8" Oct 08 06:51:53 crc kubenswrapper[4810]: E1008 06:51:53.536419 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8\": container with ID starting with a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8 not found: ID does not exist" containerID="a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.536479 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8"} err="failed to get container status \"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8\": rpc error: code = NotFound desc = could not find container \"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8\": container with ID starting with a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8 not found: ID does not exist" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.536517 4810 scope.go:117] "RemoveContainer" containerID="3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.537168 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc"} err="failed to get container status \"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc\": rpc error: code = NotFound desc = could not find container \"3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc\": container with ID starting with 3eb6fd30f63c6d611160796720a2eb6b06a60e2f6d2c405813635b8b0813c5dc not found: ID does not exist" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.537195 4810 scope.go:117] "RemoveContainer" containerID="a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.537672 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8"} err="failed to get container status \"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8\": rpc error: code = NotFound desc = could not find container \"a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8\": container with ID starting with a9f87de4b8cf6f98059bc9f202b28c114fea902d202dc06247f41db0299e63a8 not found: ID does not exist" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.558380 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "052c9e6d-bec3-4249-8c2a-742b4eea9aa1" (UID: "052c9e6d-bec3-4249-8c2a-742b4eea9aa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.579918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "052c9e6d-bec3-4249-8c2a-742b4eea9aa1" (UID: "052c9e6d-bec3-4249-8c2a-742b4eea9aa1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.585360 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "052c9e6d-bec3-4249-8c2a-742b4eea9aa1" (UID: "052c9e6d-bec3-4249-8c2a-742b4eea9aa1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.603238 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.603303 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.603317 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t58vx\" (UniqueName: \"kubernetes.io/projected/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-kube-api-access-t58vx\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.603332 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.603348 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052c9e6d-bec3-4249-8c2a-742b4eea9aa1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.860016 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.873504 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.878678 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:53 crc kubenswrapper[4810]: E1008 06:51:53.879101 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerName="init" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879117 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerName="init" Oct 08 06:51:53 crc kubenswrapper[4810]: E1008 06:51:53.879149 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-log" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879156 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-log" Oct 08 06:51:53 crc kubenswrapper[4810]: E1008 06:51:53.879168 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerName="dnsmasq-dns" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879174 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerName="dnsmasq-dns" Oct 08 06:51:53 crc kubenswrapper[4810]: E1008 06:51:53.879186 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-api" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879193 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-api" Oct 08 06:51:53 crc kubenswrapper[4810]: E1008 06:51:53.879203 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f82549-5e55-491b-9c01-bca9f29e25f3" containerName="nova-manage" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879208 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f82549-5e55-491b-9c01-bca9f29e25f3" containerName="nova-manage" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879391 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-log" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879404 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f82549-5e55-491b-9c01-bca9f29e25f3" containerName="nova-manage" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879412 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" containerName="nova-api-api" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.879426 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="89c96d63-b16e-40bb-9acd-a60a85cd8af2" containerName="dnsmasq-dns" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.880472 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.882800 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.884096 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.884184 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.894105 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:53 crc kubenswrapper[4810]: I1008 06:51:53.952016 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.015696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.016524 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqpmt\" (UniqueName: \"kubernetes.io/projected/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-kube-api-access-fqpmt\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.016581 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-logs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.016608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-public-tls-certs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.016640 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.016687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-config-data\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.083733 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="052c9e6d-bec3-4249-8c2a-742b4eea9aa1" path="/var/lib/kubelet/pods/052c9e6d-bec3-4249-8c2a-742b4eea9aa1/volumes" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.117778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-combined-ca-bundle\") pod \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.118498 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8bhn\" (UniqueName: \"kubernetes.io/projected/36e4c82b-30d2-4e56-8b1a-9646115da8b2-kube-api-access-t8bhn\") pod \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.118562 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-config-data\") pod \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\" (UID: \"36e4c82b-30d2-4e56-8b1a-9646115da8b2\") " Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.118719 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.118782 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-config-data\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.118836 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.118951 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqpmt\" (UniqueName: \"kubernetes.io/projected/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-kube-api-access-fqpmt\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.119334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-logs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.119379 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-public-tls-certs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.120746 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-logs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.122343 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36e4c82b-30d2-4e56-8b1a-9646115da8b2-kube-api-access-t8bhn" (OuterVolumeSpecName: "kube-api-access-t8bhn") pod "36e4c82b-30d2-4e56-8b1a-9646115da8b2" (UID: "36e4c82b-30d2-4e56-8b1a-9646115da8b2"). InnerVolumeSpecName "kube-api-access-t8bhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.123254 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-config-data\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.124875 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.126378 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-public-tls-certs\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.132049 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.148130 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36e4c82b-30d2-4e56-8b1a-9646115da8b2" (UID: "36e4c82b-30d2-4e56-8b1a-9646115da8b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.149353 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqpmt\" (UniqueName: \"kubernetes.io/projected/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-kube-api-access-fqpmt\") pod \"nova-api-0\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.170799 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-config-data" (OuterVolumeSpecName: "config-data") pod "36e4c82b-30d2-4e56-8b1a-9646115da8b2" (UID: "36e4c82b-30d2-4e56-8b1a-9646115da8b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.222694 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.222743 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8bhn\" (UniqueName: \"kubernetes.io/projected/36e4c82b-30d2-4e56-8b1a-9646115da8b2-kube-api-access-t8bhn\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.222759 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36e4c82b-30d2-4e56-8b1a-9646115da8b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.265755 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.420990 4810 generic.go:334] "Generic (PLEG): container finished" podID="36e4c82b-30d2-4e56-8b1a-9646115da8b2" containerID="e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7" exitCode=0 Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.421242 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.421266 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36e4c82b-30d2-4e56-8b1a-9646115da8b2","Type":"ContainerDied","Data":"e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7"} Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.421503 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"36e4c82b-30d2-4e56-8b1a-9646115da8b2","Type":"ContainerDied","Data":"c45bfc2b71d71ca54b4364c9d436851915d7670e3a336cbe6f35e7b7e652c483"} Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.421542 4810 scope.go:117] "RemoveContainer" containerID="e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.465139 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.477815 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.501114 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:54 crc kubenswrapper[4810]: E1008 06:51:54.501585 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e4c82b-30d2-4e56-8b1a-9646115da8b2" containerName="nova-scheduler-scheduler" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.501606 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e4c82b-30d2-4e56-8b1a-9646115da8b2" containerName="nova-scheduler-scheduler" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.501858 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e4c82b-30d2-4e56-8b1a-9646115da8b2" containerName="nova-scheduler-scheduler" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.502516 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.505469 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.516140 4810 scope.go:117] "RemoveContainer" containerID="e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7" Oct 08 06:51:54 crc kubenswrapper[4810]: E1008 06:51:54.518468 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7\": container with ID starting with e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7 not found: ID does not exist" containerID="e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.518501 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7"} err="failed to get container status \"e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7\": rpc error: code = NotFound desc = could not find container \"e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7\": container with ID starting with e242b79732349f22f52066901bb4eb334f840f046b01975f9045440ef34263f7 not found: ID does not exist" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.526414 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5nrg\" (UniqueName: \"kubernetes.io/projected/73271be2-0048-432e-b389-183258582254-kube-api-access-n5nrg\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.526494 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-config-data\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.526631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.527089 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.628004 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5nrg\" (UniqueName: \"kubernetes.io/projected/73271be2-0048-432e-b389-183258582254-kube-api-access-n5nrg\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.628071 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-config-data\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.628186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.632913 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.633526 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-config-data\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.646323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5nrg\" (UniqueName: \"kubernetes.io/projected/73271be2-0048-432e-b389-183258582254-kube-api-access-n5nrg\") pod \"nova-scheduler-0\" (UID: \"73271be2-0048-432e-b389-183258582254\") " pod="openstack/nova-scheduler-0" Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.752895 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:51:54 crc kubenswrapper[4810]: W1008 06:51:54.769856 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf266b088_0be1_4d57_b8f9_c891e8ad1ed3.slice/crio-ecf756d03b53ec62326985bad7be058e4218947efde29ea2cde6f0ae21d613e5 WatchSource:0}: Error finding container ecf756d03b53ec62326985bad7be058e4218947efde29ea2cde6f0ae21d613e5: Status 404 returned error can't find the container with id ecf756d03b53ec62326985bad7be058e4218947efde29ea2cde6f0ae21d613e5 Oct 08 06:51:54 crc kubenswrapper[4810]: I1008 06:51:54.826314 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:51:55 crc kubenswrapper[4810]: W1008 06:51:55.316260 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73271be2_0048_432e_b389_183258582254.slice/crio-5610b5ce3416d1a08ec85bdf6223fc5f70a0e4ee6a1bcee0583bfab9b49a6967 WatchSource:0}: Error finding container 5610b5ce3416d1a08ec85bdf6223fc5f70a0e4ee6a1bcee0583bfab9b49a6967: Status 404 returned error can't find the container with id 5610b5ce3416d1a08ec85bdf6223fc5f70a0e4ee6a1bcee0583bfab9b49a6967 Oct 08 06:51:55 crc kubenswrapper[4810]: I1008 06:51:55.319860 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:51:55 crc kubenswrapper[4810]: I1008 06:51:55.430416 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73271be2-0048-432e-b389-183258582254","Type":"ContainerStarted","Data":"5610b5ce3416d1a08ec85bdf6223fc5f70a0e4ee6a1bcee0583bfab9b49a6967"} Oct 08 06:51:55 crc kubenswrapper[4810]: I1008 06:51:55.433163 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f266b088-0be1-4d57-b8f9-c891e8ad1ed3","Type":"ContainerStarted","Data":"148c54546a54c328b2db6ab9bedc154ff074db54ed387f76af7ff0b53e27848e"} Oct 08 06:51:55 crc kubenswrapper[4810]: I1008 06:51:55.433203 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f266b088-0be1-4d57-b8f9-c891e8ad1ed3","Type":"ContainerStarted","Data":"744e6bf0d1c62016c3949d6f7346fe561790ee47cbd2b70695cea07476938ff0"} Oct 08 06:51:55 crc kubenswrapper[4810]: I1008 06:51:55.433221 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f266b088-0be1-4d57-b8f9-c891e8ad1ed3","Type":"ContainerStarted","Data":"ecf756d03b53ec62326985bad7be058e4218947efde29ea2cde6f0ae21d613e5"} Oct 08 06:51:55 crc kubenswrapper[4810]: I1008 06:51:55.458808 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.458782342 podStartE2EDuration="2.458782342s" podCreationTimestamp="2025-10-08 06:51:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:55.452451226 +0000 UTC m=+1218.086890966" watchObservedRunningTime="2025-10-08 06:51:55.458782342 +0000 UTC m=+1218.093222082" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.115332 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36e4c82b-30d2-4e56-8b1a-9646115da8b2" path="/var/lib/kubelet/pods/36e4c82b-30d2-4e56-8b1a-9646115da8b2/volumes" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.277999 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.371245 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-config-data\") pod \"244d0963-dd41-46da-89ad-8ea690f0f55d\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.371650 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244d0963-dd41-46da-89ad-8ea690f0f55d-logs\") pod \"244d0963-dd41-46da-89ad-8ea690f0f55d\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.371746 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8mpw\" (UniqueName: \"kubernetes.io/projected/244d0963-dd41-46da-89ad-8ea690f0f55d-kube-api-access-z8mpw\") pod \"244d0963-dd41-46da-89ad-8ea690f0f55d\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.371803 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-combined-ca-bundle\") pod \"244d0963-dd41-46da-89ad-8ea690f0f55d\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.371886 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-nova-metadata-tls-certs\") pod \"244d0963-dd41-46da-89ad-8ea690f0f55d\" (UID: \"244d0963-dd41-46da-89ad-8ea690f0f55d\") " Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.372271 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244d0963-dd41-46da-89ad-8ea690f0f55d-logs" (OuterVolumeSpecName: "logs") pod "244d0963-dd41-46da-89ad-8ea690f0f55d" (UID: "244d0963-dd41-46da-89ad-8ea690f0f55d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.372569 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/244d0963-dd41-46da-89ad-8ea690f0f55d-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.393111 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244d0963-dd41-46da-89ad-8ea690f0f55d-kube-api-access-z8mpw" (OuterVolumeSpecName: "kube-api-access-z8mpw") pod "244d0963-dd41-46da-89ad-8ea690f0f55d" (UID: "244d0963-dd41-46da-89ad-8ea690f0f55d"). InnerVolumeSpecName "kube-api-access-z8mpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.411921 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "244d0963-dd41-46da-89ad-8ea690f0f55d" (UID: "244d0963-dd41-46da-89ad-8ea690f0f55d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.435367 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-config-data" (OuterVolumeSpecName: "config-data") pod "244d0963-dd41-46da-89ad-8ea690f0f55d" (UID: "244d0963-dd41-46da-89ad-8ea690f0f55d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.440644 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "244d0963-dd41-46da-89ad-8ea690f0f55d" (UID: "244d0963-dd41-46da-89ad-8ea690f0f55d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.452802 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73271be2-0048-432e-b389-183258582254","Type":"ContainerStarted","Data":"4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c"} Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.458211 4810 generic.go:334] "Generic (PLEG): container finished" podID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerID="3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48" exitCode=0 Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.458795 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.460121 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"244d0963-dd41-46da-89ad-8ea690f0f55d","Type":"ContainerDied","Data":"3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48"} Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.460449 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"244d0963-dd41-46da-89ad-8ea690f0f55d","Type":"ContainerDied","Data":"d96b2c43cb06f0c7a16b0ffb962103590841f8a7bcccb70363190d8d2ba7c33b"} Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.460516 4810 scope.go:117] "RemoveContainer" containerID="3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.480503 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8mpw\" (UniqueName: \"kubernetes.io/projected/244d0963-dd41-46da-89ad-8ea690f0f55d-kube-api-access-z8mpw\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.482475 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.482571 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.482659 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/244d0963-dd41-46da-89ad-8ea690f0f55d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.506536 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.506512732 podStartE2EDuration="2.506512732s" podCreationTimestamp="2025-10-08 06:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:56.479626325 +0000 UTC m=+1219.114066065" watchObservedRunningTime="2025-10-08 06:51:56.506512732 +0000 UTC m=+1219.140952472" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.529731 4810 scope.go:117] "RemoveContainer" containerID="3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.544398 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.558562 4810 scope.go:117] "RemoveContainer" containerID="3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48" Oct 08 06:51:56 crc kubenswrapper[4810]: E1008 06:51:56.559039 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48\": container with ID starting with 3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48 not found: ID does not exist" containerID="3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.559109 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48"} err="failed to get container status \"3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48\": rpc error: code = NotFound desc = could not find container \"3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48\": container with ID starting with 3e8686c2511038fe0c978edf0e687e05a34ebd5d9e82087e84d435324496de48 not found: ID does not exist" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.559138 4810 scope.go:117] "RemoveContainer" containerID="3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633" Oct 08 06:51:56 crc kubenswrapper[4810]: E1008 06:51:56.559431 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633\": container with ID starting with 3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633 not found: ID does not exist" containerID="3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.559459 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633"} err="failed to get container status \"3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633\": rpc error: code = NotFound desc = could not find container \"3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633\": container with ID starting with 3bfaf3cda9af183a742bacfa43f84c83adcc5b0118b5ffeccd944db73f222633 not found: ID does not exist" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.560810 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.571843 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:56 crc kubenswrapper[4810]: E1008 06:51:56.572531 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-log" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.572555 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-log" Oct 08 06:51:56 crc kubenswrapper[4810]: E1008 06:51:56.572585 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-metadata" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.572594 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-metadata" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.572817 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-metadata" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.572843 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" containerName="nova-metadata-log" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.574225 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.581365 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.587514 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.587575 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.587682 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8832ab0b-b264-4db9-8b70-87920c1e4826-logs\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.587729 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qlpt\" (UniqueName: \"kubernetes.io/projected/8832ab0b-b264-4db9-8b70-87920c1e4826-kube-api-access-4qlpt\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.588085 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.588450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-config-data\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.598699 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.690476 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.690557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8832ab0b-b264-4db9-8b70-87920c1e4826-logs\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.690598 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qlpt\" (UniqueName: \"kubernetes.io/projected/8832ab0b-b264-4db9-8b70-87920c1e4826-kube-api-access-4qlpt\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.690677 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.690713 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-config-data\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.691025 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8832ab0b-b264-4db9-8b70-87920c1e4826-logs\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.696700 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.696908 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.697145 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-config-data\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.715578 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qlpt\" (UniqueName: \"kubernetes.io/projected/8832ab0b-b264-4db9-8b70-87920c1e4826-kube-api-access-4qlpt\") pod \"nova-metadata-0\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " pod="openstack/nova-metadata-0" Oct 08 06:51:56 crc kubenswrapper[4810]: I1008 06:51:56.897265 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:51:57 crc kubenswrapper[4810]: I1008 06:51:57.396046 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:51:57 crc kubenswrapper[4810]: I1008 06:51:57.477949 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8832ab0b-b264-4db9-8b70-87920c1e4826","Type":"ContainerStarted","Data":"da193473bb6de6b14dbdc891fa8dc66336bfed450ac1381353ef6ee39f88c7c0"} Oct 08 06:51:58 crc kubenswrapper[4810]: I1008 06:51:58.084790 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="244d0963-dd41-46da-89ad-8ea690f0f55d" path="/var/lib/kubelet/pods/244d0963-dd41-46da-89ad-8ea690f0f55d/volumes" Oct 08 06:51:58 crc kubenswrapper[4810]: I1008 06:51:58.494027 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8832ab0b-b264-4db9-8b70-87920c1e4826","Type":"ContainerStarted","Data":"c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4"} Oct 08 06:51:58 crc kubenswrapper[4810]: I1008 06:51:58.494099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8832ab0b-b264-4db9-8b70-87920c1e4826","Type":"ContainerStarted","Data":"4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84"} Oct 08 06:51:58 crc kubenswrapper[4810]: I1008 06:51:58.530158 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.530126038 podStartE2EDuration="2.530126038s" podCreationTimestamp="2025-10-08 06:51:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 06:51:58.519343119 +0000 UTC m=+1221.153782849" watchObservedRunningTime="2025-10-08 06:51:58.530126038 +0000 UTC m=+1221.164565818" Oct 08 06:51:59 crc kubenswrapper[4810]: I1008 06:51:59.826932 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 06:52:01 crc kubenswrapper[4810]: I1008 06:52:01.897615 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 06:52:01 crc kubenswrapper[4810]: I1008 06:52:01.897994 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 06:52:04 crc kubenswrapper[4810]: I1008 06:52:04.266168 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 06:52:04 crc kubenswrapper[4810]: I1008 06:52:04.266562 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 06:52:04 crc kubenswrapper[4810]: I1008 06:52:04.827270 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 06:52:04 crc kubenswrapper[4810]: I1008 06:52:04.858057 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 06:52:05 crc kubenswrapper[4810]: I1008 06:52:05.278183 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 06:52:05 crc kubenswrapper[4810]: I1008 06:52:05.278183 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 06:52:05 crc kubenswrapper[4810]: I1008 06:52:05.626656 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 06:52:06 crc kubenswrapper[4810]: I1008 06:52:06.898206 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 06:52:06 crc kubenswrapper[4810]: I1008 06:52:06.898704 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 06:52:07 crc kubenswrapper[4810]: I1008 06:52:07.912114 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 06:52:07 crc kubenswrapper[4810]: I1008 06:52:07.913174 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 06:52:12 crc kubenswrapper[4810]: I1008 06:52:12.941323 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 06:52:14 crc kubenswrapper[4810]: I1008 06:52:14.275736 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 06:52:14 crc kubenswrapper[4810]: I1008 06:52:14.276137 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 06:52:14 crc kubenswrapper[4810]: I1008 06:52:14.276381 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 06:52:14 crc kubenswrapper[4810]: I1008 06:52:14.276410 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 06:52:14 crc kubenswrapper[4810]: I1008 06:52:14.285556 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 06:52:14 crc kubenswrapper[4810]: I1008 06:52:14.286154 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 06:52:16 crc kubenswrapper[4810]: I1008 06:52:16.906271 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 06:52:16 crc kubenswrapper[4810]: I1008 06:52:16.906621 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 06:52:16 crc kubenswrapper[4810]: I1008 06:52:16.916079 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 06:52:16 crc kubenswrapper[4810]: I1008 06:52:16.916407 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.489698 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.490431 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" containerName="openstackclient" containerID="cri-o://a1c4944dcfd55a27b1f280aa67cbd8b54e526e0b2822d8a76c6d1e57a1032b88" gracePeriod=2 Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.503859 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.889400 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementc68e-account-delete-m9f52"] Oct 08 06:52:34 crc kubenswrapper[4810]: E1008 06:52:34.890021 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" containerName="openstackclient" Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.890100 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" containerName="openstackclient" Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.890340 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" containerName="openstackclient" Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.891015 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementc68e-account-delete-m9f52" Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.910693 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementc68e-account-delete-m9f52"] Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.946183 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.993071 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder95b0-account-delete-l4qbw"] Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.994168 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrp8z\" (UniqueName: \"kubernetes.io/projected/ced62b16-201e-4f75-a970-20bad14e1441-kube-api-access-rrp8z\") pod \"placementc68e-account-delete-m9f52\" (UID: \"ced62b16-201e-4f75-a970-20bad14e1441\") " pod="openstack/placementc68e-account-delete-m9f52" Oct 08 06:52:34 crc kubenswrapper[4810]: I1008 06:52:34.994538 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder95b0-account-delete-l4qbw" Oct 08 06:52:34 crc kubenswrapper[4810]: E1008 06:52:34.994691 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 06:52:34 crc kubenswrapper[4810]: E1008 06:52:34.994738 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data podName:2a44c96c-b96c-409c-9c9e-c049d9fe68b5 nodeName:}" failed. No retries permitted until 2025-10-08 06:52:35.494722656 +0000 UTC m=+1258.129162396 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data") pod "rabbitmq-server-0" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5") : configmap "rabbitmq-config-data" not found Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.033028 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder95b0-account-delete-l4qbw"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.095454 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrp8z\" (UniqueName: \"kubernetes.io/projected/ced62b16-201e-4f75-a970-20bad14e1441-kube-api-access-rrp8z\") pod \"placementc68e-account-delete-m9f52\" (UID: \"ced62b16-201e-4f75-a970-20bad14e1441\") " pod="openstack/placementc68e-account-delete-m9f52" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.099031 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rb7\" (UniqueName: \"kubernetes.io/projected/90da3b83-b1b7-4d11-9f4d-37cf76b93edb-kube-api-access-85rb7\") pod \"cinder95b0-account-delete-l4qbw\" (UID: \"90da3b83-b1b7-4d11-9f4d-37cf76b93edb\") " pod="openstack/cinder95b0-account-delete-l4qbw" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.130635 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrp8z\" (UniqueName: \"kubernetes.io/projected/ced62b16-201e-4f75-a970-20bad14e1441-kube-api-access-rrp8z\") pod \"placementc68e-account-delete-m9f52\" (UID: \"ced62b16-201e-4f75-a970-20bad14e1441\") " pod="openstack/placementc68e-account-delete-m9f52" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.151891 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.196950 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronb73c-account-delete-7cpkv"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.198466 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb73c-account-delete-7cpkv" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.201178 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rb7\" (UniqueName: \"kubernetes.io/projected/90da3b83-b1b7-4d11-9f4d-37cf76b93edb-kube-api-access-85rb7\") pod \"cinder95b0-account-delete-l4qbw\" (UID: \"90da3b83-b1b7-4d11-9f4d-37cf76b93edb\") " pod="openstack/cinder95b0-account-delete-l4qbw" Oct 08 06:52:35 crc kubenswrapper[4810]: E1008 06:52:35.202068 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:35 crc kubenswrapper[4810]: E1008 06:52:35.202752 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data podName:4deb0612-547f-4067-b95b-5794663d21aa nodeName:}" failed. No retries permitted until 2025-10-08 06:52:35.702721936 +0000 UTC m=+1258.337161676 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data") pod "rabbitmq-cell1-server-0" (UID: "4deb0612-547f-4067-b95b-5794663d21aa") : configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.215268 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.215985 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="openstack-network-exporter" containerID="cri-o://02937c8f188e832e3f815057f823449d155b5ae31271a2816bdd2a5efb1d99a8" gracePeriod=300 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.226288 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementc68e-account-delete-m9f52" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.233814 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronb73c-account-delete-7cpkv"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.237825 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rb7\" (UniqueName: \"kubernetes.io/projected/90da3b83-b1b7-4d11-9f4d-37cf76b93edb-kube-api-access-85rb7\") pod \"cinder95b0-account-delete-l4qbw\" (UID: \"90da3b83-b1b7-4d11-9f4d-37cf76b93edb\") " pod="openstack/cinder95b0-account-delete-l4qbw" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.286024 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-h7z72"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.296114 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-h7z72"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.308785 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl2fl\" (UniqueName: \"kubernetes.io/projected/9c1351c1-21aa-4004-a402-a603343c58d7-kube-api-access-fl2fl\") pod \"neutronb73c-account-delete-7cpkv\" (UID: \"9c1351c1-21aa-4004-a402-a603343c58d7\") " pod="openstack/neutronb73c-account-delete-7cpkv" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.319082 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.319348 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="ovn-northd" containerID="cri-o://a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" gracePeriod=30 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.320163 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="openstack-network-exporter" containerID="cri-o://6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b" gracePeriod=30 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.323486 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder95b0-account-delete-l4qbw" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.367712 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="ovsdbserver-nb" containerID="cri-o://e4123f776f7b7b29b557fdc6e108f0d0848f10b0a0ef5e629afe696812883bbf" gracePeriod=300 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.395038 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi29ac-account-delete-mjvf2"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.396593 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi29ac-account-delete-mjvf2" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.408780 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi29ac-account-delete-mjvf2"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.411202 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl2fl\" (UniqueName: \"kubernetes.io/projected/9c1351c1-21aa-4004-a402-a603343c58d7-kube-api-access-fl2fl\") pod \"neutronb73c-account-delete-7cpkv\" (UID: \"9c1351c1-21aa-4004-a402-a603343c58d7\") " pod="openstack/neutronb73c-account-delete-7cpkv" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.465824 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl2fl\" (UniqueName: \"kubernetes.io/projected/9c1351c1-21aa-4004-a402-a603343c58d7-kube-api-access-fl2fl\") pod \"neutronb73c-account-delete-7cpkv\" (UID: \"9c1351c1-21aa-4004-a402-a603343c58d7\") " pod="openstack/neutronb73c-account-delete-7cpkv" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.519534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb73c-account-delete-7cpkv" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.521910 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gm5l\" (UniqueName: \"kubernetes.io/projected/b7b1ed43-571a-4b70-9eae-c19a7675bc59-kube-api-access-4gm5l\") pod \"novaapi29ac-account-delete-mjvf2\" (UID: \"b7b1ed43-571a-4b70-9eae-c19a7675bc59\") " pod="openstack/novaapi29ac-account-delete-mjvf2" Oct 08 06:52:35 crc kubenswrapper[4810]: E1008 06:52:35.522194 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 06:52:35 crc kubenswrapper[4810]: E1008 06:52:35.522243 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data podName:2a44c96c-b96c-409c-9c9e-c049d9fe68b5 nodeName:}" failed. No retries permitted until 2025-10-08 06:52:36.522230206 +0000 UTC m=+1259.156669946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data") pod "rabbitmq-server-0" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5") : configmap "rabbitmq-config-data" not found Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.527863 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.528464 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="openstack-network-exporter" containerID="cri-o://b2cbef70cc311c4f05f44991c29903743960b2ccd17d6f94057bd5df3a186bbb" gracePeriod=300 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.550584 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-qtj9x"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.605302 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-xlsjm"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.628558 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gm5l\" (UniqueName: \"kubernetes.io/projected/b7b1ed43-571a-4b70-9eae-c19a7675bc59-kube-api-access-4gm5l\") pod \"novaapi29ac-account-delete-mjvf2\" (UID: \"b7b1ed43-571a-4b70-9eae-c19a7675bc59\") " pod="openstack/novaapi29ac-account-delete-mjvf2" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.682001 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0627c-account-delete-z6g9h"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.683312 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0627c-account-delete-z6g9h" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.723801 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gm5l\" (UniqueName: \"kubernetes.io/projected/b7b1ed43-571a-4b70-9eae-c19a7675bc59-kube-api-access-4gm5l\") pod \"novaapi29ac-account-delete-mjvf2\" (UID: \"b7b1ed43-571a-4b70-9eae-c19a7675bc59\") " pod="openstack/novaapi29ac-account-delete-mjvf2" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.731159 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrfhz\" (UniqueName: \"kubernetes.io/projected/64350cd1-38b8-4c64-8c01-105432d1b91f-kube-api-access-jrfhz\") pod \"novacell0627c-account-delete-z6g9h\" (UID: \"64350cd1-38b8-4c64-8c01-105432d1b91f\") " pod="openstack/novacell0627c-account-delete-z6g9h" Oct 08 06:52:35 crc kubenswrapper[4810]: E1008 06:52:35.731492 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:35 crc kubenswrapper[4810]: E1008 06:52:35.731541 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data podName:4deb0612-547f-4067-b95b-5794663d21aa nodeName:}" failed. No retries permitted until 2025-10-08 06:52:36.731524252 +0000 UTC m=+1259.365963992 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data") pod "rabbitmq-cell1-server-0" (UID: "4deb0612-547f-4067-b95b-5794663d21aa") : configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.748245 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-wqx77"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.756852 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="ovsdbserver-sb" containerID="cri-o://9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6" gracePeriod=300 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.757267 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0627c-account-delete-z6g9h"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.768797 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-wqx77"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.817162 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fchxc"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.835181 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-fchxc" podUID="d6f5fbdd-0691-4ce8-888d-be0147974607" containerName="openstack-network-exporter" containerID="cri-o://8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49" gracePeriod=30 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.841354 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi29ac-account-delete-mjvf2" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.863731 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrfhz\" (UniqueName: \"kubernetes.io/projected/64350cd1-38b8-4c64-8c01-105432d1b91f-kube-api-access-jrfhz\") pod \"novacell0627c-account-delete-z6g9h\" (UID: \"64350cd1-38b8-4c64-8c01-105432d1b91f\") " pod="openstack/novacell0627c-account-delete-z6g9h" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.895785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrfhz\" (UniqueName: \"kubernetes.io/projected/64350cd1-38b8-4c64-8c01-105432d1b91f-kube-api-access-jrfhz\") pod \"novacell0627c-account-delete-z6g9h\" (UID: \"64350cd1-38b8-4c64-8c01-105432d1b91f\") " pod="openstack/novacell0627c-account-delete-z6g9h" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.896112 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-lcpxn"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.928576 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-lcpxn"] Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.947048 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4a3bceca-c9a2-45e1-9f81-2bd9f6985c36/ovsdbserver-nb/0.log" Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.947094 4810 generic.go:334] "Generic (PLEG): container finished" podID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerID="02937c8f188e832e3f815057f823449d155b5ae31271a2816bdd2a5efb1d99a8" exitCode=2 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.947110 4810 generic.go:334] "Generic (PLEG): container finished" podID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerID="e4123f776f7b7b29b557fdc6e108f0d0848f10b0a0ef5e629afe696812883bbf" exitCode=143 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.947166 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36","Type":"ContainerDied","Data":"02937c8f188e832e3f815057f823449d155b5ae31271a2816bdd2a5efb1d99a8"} Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.947192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36","Type":"ContainerDied","Data":"e4123f776f7b7b29b557fdc6e108f0d0848f10b0a0ef5e629afe696812883bbf"} Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.993217 4810 generic.go:334] "Generic (PLEG): container finished" podID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerID="b2cbef70cc311c4f05f44991c29903743960b2ccd17d6f94057bd5df3a186bbb" exitCode=2 Oct 08 06:52:35 crc kubenswrapper[4810]: I1008 06:52:35.993303 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a9ba387c-c5ce-4f6e-8323-7e6afce894ba","Type":"ContainerDied","Data":"b2cbef70cc311c4f05f44991c29903743960b2ccd17d6f94057bd5df3a186bbb"} Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.014854 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0627c-account-delete-z6g9h" Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.018426 4810 generic.go:334] "Generic (PLEG): container finished" podID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerID="6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b" exitCode=2 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.018502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7","Type":"ContainerDied","Data":"6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b"} Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.031672 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-7hkxc"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.055300 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-7hkxc"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.127115 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22742881-9836-4dbb-bd8c-deba4194d7c3" path="/var/lib/kubelet/pods/22742881-9836-4dbb-bd8c-deba4194d7c3/volumes" Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.142852 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82b5e634-87c4-4b75-9d3d-b78816226084" path="/var/lib/kubelet/pods/82b5e634-87c4-4b75-9d3d-b78816226084/volumes" Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.154361 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c40f9ad6-fed6-4759-8aa3-134912fd46bc" path="/var/lib/kubelet/pods/c40f9ad6-fed6-4759-8aa3-134912fd46bc/volumes" Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.155563 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb95c243-0c64-4937-a86e-34f348023513" path="/var/lib/kubelet/pods/cb95c243-0c64-4937-a86e-34f348023513/volumes" Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.156946 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d4d96bb9-89r5m"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.157361 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" podUID="5091c146-f092-4878-95a3-49cb91760d17" containerName="dnsmasq-dns" containerID="cri-o://7d7302d109e2280340776457b1d206311a58db6e8a68f517aea3acb2a3267097" gracePeriod=10 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.169441 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-bx46b"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.192424 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-bx46b"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.284690 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-684c4d4db6-dr9wq"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.285131 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-684c4d4db6-dr9wq" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-log" containerID="cri-o://3556449533e62b8a1b045f6331743feae434169e2723d25a19669ad4c27722e9" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.285552 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-684c4d4db6-dr9wq" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-api" containerID="cri-o://b526413b9a8c8e6cae317ae5e79f9dced75f63320d48b38b7a09dab4a2bdcb30" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.323037 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9knv8"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.367188 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9knv8"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.410480 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-vfmcm"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.453554 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-vfmcm"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.460926 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementc68e-account-delete-m9f52"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.471179 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-trvz6"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.479769 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-trvz6"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.494422 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.494907 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api-log" containerID="cri-o://811cbf9580f361194bc54a2a641dd527653e175eccb2b963cfa94e4f09a79794" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.495306 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api" containerID="cri-o://8d3bf6f25cf1c45bf48dbb9c1a31f7c04f185f3159a9935da5533577297101f4" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.509136 4810 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 08 06:52:36 crc kubenswrapper[4810]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 08 06:52:36 crc kubenswrapper[4810]: + source /usr/local/bin/container-scripts/functions Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNBridge=br-int Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNRemote=tcp:localhost:6642 Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNEncapType=geneve Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNAvailabilityZones= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ EnableChassisAsGateway=true Oct 08 06:52:36 crc kubenswrapper[4810]: ++ PhysicalNetworks= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNHostName= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 08 06:52:36 crc kubenswrapper[4810]: ++ ovs_dir=/var/lib/openvswitch Oct 08 06:52:36 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 08 06:52:36 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 08 06:52:36 crc kubenswrapper[4810]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 06:52:36 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 06:52:36 crc kubenswrapper[4810]: + sleep 0.5 Oct 08 06:52:36 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 06:52:36 crc kubenswrapper[4810]: + cleanup_ovsdb_server_semaphore Oct 08 06:52:36 crc kubenswrapper[4810]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 06:52:36 crc kubenswrapper[4810]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 08 06:52:36 crc kubenswrapper[4810]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-xlsjm" message=< Oct 08 06:52:36 crc kubenswrapper[4810]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 08 06:52:36 crc kubenswrapper[4810]: + source /usr/local/bin/container-scripts/functions Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNBridge=br-int Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNRemote=tcp:localhost:6642 Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNEncapType=geneve Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNAvailabilityZones= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ EnableChassisAsGateway=true Oct 08 06:52:36 crc kubenswrapper[4810]: ++ PhysicalNetworks= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNHostName= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 08 06:52:36 crc kubenswrapper[4810]: ++ ovs_dir=/var/lib/openvswitch Oct 08 06:52:36 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 08 06:52:36 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 08 06:52:36 crc kubenswrapper[4810]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 06:52:36 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 06:52:36 crc kubenswrapper[4810]: + sleep 0.5 Oct 08 06:52:36 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 06:52:36 crc kubenswrapper[4810]: + cleanup_ovsdb_server_semaphore Oct 08 06:52:36 crc kubenswrapper[4810]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 06:52:36 crc kubenswrapper[4810]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 08 06:52:36 crc kubenswrapper[4810]: > Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.509190 4810 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 08 06:52:36 crc kubenswrapper[4810]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 08 06:52:36 crc kubenswrapper[4810]: + source /usr/local/bin/container-scripts/functions Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNBridge=br-int Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNRemote=tcp:localhost:6642 Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNEncapType=geneve Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNAvailabilityZones= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ EnableChassisAsGateway=true Oct 08 06:52:36 crc kubenswrapper[4810]: ++ PhysicalNetworks= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ OVNHostName= Oct 08 06:52:36 crc kubenswrapper[4810]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 08 06:52:36 crc kubenswrapper[4810]: ++ ovs_dir=/var/lib/openvswitch Oct 08 06:52:36 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 08 06:52:36 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 08 06:52:36 crc kubenswrapper[4810]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 06:52:36 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 06:52:36 crc kubenswrapper[4810]: + sleep 0.5 Oct 08 06:52:36 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 06:52:36 crc kubenswrapper[4810]: + cleanup_ovsdb_server_semaphore Oct 08 06:52:36 crc kubenswrapper[4810]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 06:52:36 crc kubenswrapper[4810]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 08 06:52:36 crc kubenswrapper[4810]: > pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" containerID="cri-o://f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.509226 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" containerID="cri-o://f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.515084 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.515468 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-log" containerID="cri-o://54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.516046 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-httpd" containerID="cri-o://26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.525758 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.528575 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-server" containerID="cri-o://39e49e136cc29f048dd01f7cdb829f4235eb528d5956a0e907b4abc94d91b6bd" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.528885 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="swift-recon-cron" containerID="cri-o://758c08848dcf56803d3340df9307815df8fed52c268f9d576bd87dce38f15071" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.528922 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="rsync" containerID="cri-o://b0eb359ab07ae0f5e3fe4ee340568b60987dd5232e6f155a7a05a079ef8c5e07" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.528999 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-expirer" containerID="cri-o://345bb23b7033db196c378bf859a6c3b70c55475aa015ae3af5ecd4503e2dcf38" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529036 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-updater" containerID="cri-o://600f7e0d7b7425e861bb2335da4e0b3bf0e9e3610e33eb0340d0e575eba425da" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529067 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-auditor" containerID="cri-o://f46dd8e3225e4f46200299b022317c9b2569c98b96714c24a69495ae35120418" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529096 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-replicator" containerID="cri-o://e5382afb7813f143406f3f3d996682bc6f5be13125adec6966b7d947409db0b5" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529133 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-server" containerID="cri-o://d8f77941e09530d2318f1aa89644df82f0915aa3cfe954e93ebcfc8a0be63897" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529277 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-updater" containerID="cri-o://b1c09d7235386cf593d3456b3d746d6d1cf962a7a681c42f5eb436bbfa5c7423" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529338 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-auditor" containerID="cri-o://aeceb614fb9899cfafb9c3fa078fce2b620e618a9756ff3d62a4a160e354a0b2" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529373 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-replicator" containerID="cri-o://8c66da6cd63ffdac911d7e4cb28fc214ac4f2ca0696b970109e6193d96655d61" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529412 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-server" containerID="cri-o://ce160cdae1ce3f85d7b348919fd8ec1e63ee09bafdd24ff6e4c23db7ad5a2ff1" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529447 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-reaper" containerID="cri-o://f2a45c96c2a4d65f22eb75f3b31681ceb8d4be034c534c71d36192573c688352" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529491 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-auditor" containerID="cri-o://cc83732eb2aa89586c96fc2f8b3a8d349fb000018c5d29cd6d05bd9035d6901c" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.529521 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-replicator" containerID="cri-o://f97fcf1457257c718d7bc50e69bb8e879b0bcdda1152aafd198dc4eacc3764c1" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.555112 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.555539 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-log" containerID="cri-o://47e85962732a13a5f5460d0367ec3af0533b5a5617841d2c9fda2a5452a2fbdb" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.555790 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-httpd" containerID="cri-o://ed04a8db57a733ab9675e113837de747b29cbd4384df9603ebc90167bf592eca" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.577589 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6 is running failed: container process not found" containerID="9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.580441 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6 is running failed: container process not found" containerID="9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.580863 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6 is running failed: container process not found" containerID="9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.580929 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="ovsdbserver-sb" Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.625577 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.625869 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data podName:2a44c96c-b96c-409c-9c9e-c049d9fe68b5 nodeName:}" failed. No retries permitted until 2025-10-08 06:52:38.625851561 +0000 UTC m=+1261.260291301 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data") pod "rabbitmq-server-0" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5") : configmap "rabbitmq-config-data" not found Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.653227 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" containerID="cri-o://785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" gracePeriod=29 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.658058 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.658347 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="cinder-scheduler" containerID="cri-o://f6b4a8db2b3498b1430bdc164af3693743a79260b3c2097cdb4bee62cab40e39" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.659009 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="probe" containerID="cri-o://81749551b081df35eaa781dfa19fa8b4dcb1258c7b85162b7bf3626674f4ace6" gracePeriod=30 Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.691271 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-jx9km"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.716435 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-jx9km"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.727744 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f3d6-account-create-n49kn"] Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.737892 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:36 crc kubenswrapper[4810]: E1008 06:52:36.737957 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data podName:4deb0612-547f-4067-b95b-5794663d21aa nodeName:}" failed. No retries permitted until 2025-10-08 06:52:38.737943348 +0000 UTC m=+1261.372383088 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data") pod "rabbitmq-cell1-server-0" (UID: "4deb0612-547f-4067-b95b-5794663d21aa") : configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.749268 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f3d6-account-create-n49kn"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.783152 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c68e-account-create-rz9v2"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.830236 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c68e-account-create-rz9v2"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.843373 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-hx4fs"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.865040 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-g4kvt"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.880701 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementc68e-account-delete-m9f52"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.895940 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-hx4fs"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.908113 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1485-account-create-687t8"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.934019 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-g4kvt"] Oct 08 06:52:36 crc kubenswrapper[4810]: I1008 06:52:36.943012 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1485-account-create-687t8"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.058933 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.064852 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fchxc_d6f5fbdd-0691-4ce8-888d-be0147974607/openstack-network-exporter/0.log" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.064920 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.069079 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-h68hp"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.073696 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-h68hp"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.079937 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-95b0-account-create-txvcv"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.082112 4810 generic.go:334] "Generic (PLEG): container finished" podID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerID="3556449533e62b8a1b045f6331743feae434169e2723d25a19669ad4c27722e9" exitCode=143 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.082219 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684c4d4db6-dr9wq" event={"ID":"d7486bd0-d5fd-4d4f-8596-705cc22237d5","Type":"ContainerDied","Data":"3556449533e62b8a1b045f6331743feae434169e2723d25a19669ad4c27722e9"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.090802 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6f65ff9-7hhz2"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.091120 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6f65ff9-7hhz2" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-api" containerID="cri-o://77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.091202 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a9ba387c-c5ce-4f6e-8323-7e6afce894ba/ovsdbserver-sb/0.log" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.091330 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.091459 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6f65ff9-7hhz2" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-httpd" containerID="cri-o://a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.094235 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4a3bceca-c9a2-45e1-9f81-2bd9f6985c36/ovsdbserver-nb/0.log" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.094370 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.106462 4810 generic.go:334] "Generic (PLEG): container finished" podID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerID="54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d" exitCode=143 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.106539 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dced1083-d856-4f53-bb6e-19fc64ba95a5","Type":"ContainerDied","Data":"54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.125656 4810 generic.go:334] "Generic (PLEG): container finished" podID="35e33182-d302-417e-8e4e-fec4808047a4" containerID="811cbf9580f361194bc54a2a641dd527653e175eccb2b963cfa94e4f09a79794" exitCode=143 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.125735 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"35e33182-d302-417e-8e4e-fec4808047a4","Type":"ContainerDied","Data":"811cbf9580f361194bc54a2a641dd527653e175eccb2b963cfa94e4f09a79794"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.132256 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc68e-account-delete-m9f52" event={"ID":"ced62b16-201e-4f75-a970-20bad14e1441","Type":"ContainerStarted","Data":"c791d0a3d7ca0fde0848db8aef544a00cf85490bb89d1e9e5985a4a3259cbdd7"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.134257 4810 generic.go:334] "Generic (PLEG): container finished" podID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerID="47e85962732a13a5f5460d0367ec3af0533b5a5617841d2c9fda2a5452a2fbdb" exitCode=143 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.134314 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48ca6de6-470e-4226-b67b-e58bf64415e3","Type":"ContainerDied","Data":"47e85962732a13a5f5460d0367ec3af0533b5a5617841d2c9fda2a5452a2fbdb"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.136649 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-fchxc_d6f5fbdd-0691-4ce8-888d-be0147974607/openstack-network-exporter/0.log" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.136694 4810 generic.go:334] "Generic (PLEG): container finished" podID="d6f5fbdd-0691-4ce8-888d-be0147974607" containerID="8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49" exitCode=2 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.136736 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fchxc" event={"ID":"d6f5fbdd-0691-4ce8-888d-be0147974607","Type":"ContainerDied","Data":"8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.136801 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fchxc" event={"ID":"d6f5fbdd-0691-4ce8-888d-be0147974607","Type":"ContainerDied","Data":"2c85a732a55ba049badae5cd7192c567574150bda1f0d37daccf3ad3ec2b4834"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.136824 4810 scope.go:117] "RemoveContainer" containerID="8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.137023 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fchxc" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.158094 4810 generic.go:334] "Generic (PLEG): container finished" podID="31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" containerID="a1c4944dcfd55a27b1f280aa67cbd8b54e526e0b2822d8a76c6d1e57a1032b88" exitCode=137 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.204859 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder95b0-account-delete-l4qbw"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.216337 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.216438 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xlsjm" event={"ID":"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37","Type":"ContainerDied","Data":"f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.253152 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgvsq\" (UniqueName: \"kubernetes.io/projected/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-kube-api-access-lgvsq\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268685 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-scripts\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268731 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-config\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268754 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268772 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovs-rundir\") pod \"d6f5fbdd-0691-4ce8-888d-be0147974607\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268814 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268869 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-combined-ca-bundle\") pod \"d6f5fbdd-0691-4ce8-888d-be0147974607\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268889 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-metrics-certs-tls-certs\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.268908 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-combined-ca-bundle\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269027 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdb-rundir\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269048 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-combined-ca-bundle\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269072 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-metrics-certs-tls-certs\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269086 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f5fbdd-0691-4ce8-888d-be0147974607-config\") pod \"d6f5fbdd-0691-4ce8-888d-be0147974607\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269075 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "d6f5fbdd-0691-4ce8-888d-be0147974607" (UID: "d6f5fbdd-0691-4ce8-888d-be0147974607"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269112 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdb-rundir\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269158 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-config\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269186 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdbserver-nb-tls-certs\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269220 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mblm2\" (UniqueName: \"kubernetes.io/projected/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-kube-api-access-mblm2\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdbserver-sb-tls-certs\") pod \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\" (UID: \"a9ba387c-c5ce-4f6e-8323-7e6afce894ba\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269290 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovn-rundir\") pod \"d6f5fbdd-0691-4ce8-888d-be0147974607\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269313 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-scripts\") pod \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\" (UID: \"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-metrics-certs-tls-certs\") pod \"d6f5fbdd-0691-4ce8-888d-be0147974607\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.269352 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc2pw\" (UniqueName: \"kubernetes.io/projected/d6f5fbdd-0691-4ce8-888d-be0147974607-kube-api-access-qc2pw\") pod \"d6f5fbdd-0691-4ce8-888d-be0147974607\" (UID: \"d6f5fbdd-0691-4ce8-888d-be0147974607\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.270128 4810 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.270156 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-scripts" (OuterVolumeSpecName: "scripts") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.270839 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-config" (OuterVolumeSpecName: "config") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.271542 4810 generic.go:334] "Generic (PLEG): container finished" podID="5091c146-f092-4878-95a3-49cb91760d17" containerID="7d7302d109e2280340776457b1d206311a58db6e8a68f517aea3acb2a3267097" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.271620 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" event={"ID":"5091c146-f092-4878-95a3-49cb91760d17","Type":"ContainerDied","Data":"7d7302d109e2280340776457b1d206311a58db6e8a68f517aea3acb2a3267097"} Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.271672 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.272005 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "d6f5fbdd-0691-4ce8-888d-be0147974607" (UID: "d6f5fbdd-0691-4ce8-888d-be0147974607"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.272358 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-scripts" (OuterVolumeSpecName: "scripts") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.278679 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.280356 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.280402 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.283918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.284521 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-config" (OuterVolumeSpecName: "config") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.288082 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-kube-api-access-lgvsq" (OuterVolumeSpecName: "kube-api-access-lgvsq") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "kube-api-access-lgvsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.288411 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.290294 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f5fbdd-0691-4ce8-888d-be0147974607-config" (OuterVolumeSpecName: "config") pod "d6f5fbdd-0691-4ce8-888d-be0147974607" (UID: "d6f5fbdd-0691-4ce8-888d-be0147974607"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.290386 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.303218 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.303687 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a9ba387c-c5ce-4f6e-8323-7e6afce894ba/ovsdbserver-sb/0.log" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.303729 4810 generic.go:334] "Generic (PLEG): container finished" podID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerID="9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6" exitCode=143 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.303784 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a9ba387c-c5ce-4f6e-8323-7e6afce894ba","Type":"ContainerDied","Data":"9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6"} Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.303829 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.303933 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.304947 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.316197 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-kube-api-access-mblm2" (OuterVolumeSpecName: "kube-api-access-mblm2") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "kube-api-access-mblm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.316343 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.316394 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="ovn-northd" Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.316708 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.316733 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.319183 4810 scope.go:117] "RemoveContainer" containerID="8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.334445 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: E1008 06:52:37.337720 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49\": container with ID starting with 8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49 not found: ID does not exist" containerID="8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.337764 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49"} err="failed to get container status \"8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49\": rpc error: code = NotFound desc = could not find container \"8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49\": container with ID starting with 8c4d64501b9a546fb73d0205ae97add5cc920374436bb4834fc93fcaa675be49 not found: ID does not exist" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.337795 4810 scope.go:117] "RemoveContainer" containerID="b2cbef70cc311c4f05f44991c29903743960b2ccd17d6f94057bd5df3a186bbb" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.341165 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.347165 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-95b0-account-create-txvcv"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.375888 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f5fbdd-0691-4ce8-888d-be0147974607-kube-api-access-qc2pw" (OuterVolumeSpecName: "kube-api-access-qc2pw") pod "d6f5fbdd-0691-4ce8-888d-be0147974607" (UID: "d6f5fbdd-0691-4ce8-888d-be0147974607"). InnerVolumeSpecName "kube-api-access-qc2pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377480 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgvsq\" (UniqueName: \"kubernetes.io/projected/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-kube-api-access-lgvsq\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377505 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377516 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377539 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377554 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377562 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377572 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f5fbdd-0691-4ce8-888d-be0147974607-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377581 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377589 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377596 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mblm2\" (UniqueName: \"kubernetes.io/projected/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-kube-api-access-mblm2\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377605 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d6f5fbdd-0691-4ce8-888d-be0147974607-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377613 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.377621 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc2pw\" (UniqueName: \"kubernetes.io/projected/d6f5fbdd-0691-4ce8-888d-be0147974607-kube-api-access-qc2pw\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.391363 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="rabbitmq" containerID="cri-o://5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98" gracePeriod=604800 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.437915 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-4r7z7"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470343 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="b0eb359ab07ae0f5e3fe4ee340568b60987dd5232e6f155a7a05a079ef8c5e07" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470387 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="345bb23b7033db196c378bf859a6c3b70c55475aa015ae3af5ecd4503e2dcf38" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470396 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="600f7e0d7b7425e861bb2335da4e0b3bf0e9e3610e33eb0340d0e575eba425da" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470405 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="f46dd8e3225e4f46200299b022317c9b2569c98b96714c24a69495ae35120418" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470413 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="e5382afb7813f143406f3f3d996682bc6f5be13125adec6966b7d947409db0b5" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470421 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="b1c09d7235386cf593d3456b3d746d6d1cf962a7a681c42f5eb436bbfa5c7423" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470429 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="aeceb614fb9899cfafb9c3fa078fce2b620e618a9756ff3d62a4a160e354a0b2" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470437 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="8c66da6cd63ffdac911d7e4cb28fc214ac4f2ca0696b970109e6193d96655d61" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470445 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="f2a45c96c2a4d65f22eb75f3b31681ceb8d4be034c534c71d36192573c688352" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470453 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="cc83732eb2aa89586c96fc2f8b3a8d349fb000018c5d29cd6d05bd9035d6901c" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470460 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="f97fcf1457257c718d7bc50e69bb8e879b0bcdda1152aafd198dc4eacc3764c1" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470471 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="39e49e136cc29f048dd01f7cdb829f4235eb528d5956a0e907b4abc94d91b6bd" exitCode=0 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470500 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"b0eb359ab07ae0f5e3fe4ee340568b60987dd5232e6f155a7a05a079ef8c5e07"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"345bb23b7033db196c378bf859a6c3b70c55475aa015ae3af5ecd4503e2dcf38"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470546 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"600f7e0d7b7425e861bb2335da4e0b3bf0e9e3610e33eb0340d0e575eba425da"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470557 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"f46dd8e3225e4f46200299b022317c9b2569c98b96714c24a69495ae35120418"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470568 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"e5382afb7813f143406f3f3d996682bc6f5be13125adec6966b7d947409db0b5"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470579 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"b1c09d7235386cf593d3456b3d746d6d1cf962a7a681c42f5eb436bbfa5c7423"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470590 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"aeceb614fb9899cfafb9c3fa078fce2b620e618a9756ff3d62a4a160e354a0b2"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470601 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"8c66da6cd63ffdac911d7e4cb28fc214ac4f2ca0696b970109e6193d96655d61"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470613 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"f2a45c96c2a4d65f22eb75f3b31681ceb8d4be034c534c71d36192573c688352"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470623 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"cc83732eb2aa89586c96fc2f8b3a8d349fb000018c5d29cd6d05bd9035d6901c"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470634 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"f97fcf1457257c718d7bc50e69bb8e879b0bcdda1152aafd198dc4eacc3764c1"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.470646 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"39e49e136cc29f048dd01f7cdb829f4235eb528d5956a0e907b4abc94d91b6bd"} Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.474955 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-4r7z7"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.486364 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b73c-account-create-6nkpk"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.491256 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b73c-account-create-6nkpk"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.514221 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb73c-account-delete-7cpkv"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.541029 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.551252 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.551522 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-log" containerID="cri-o://744e6bf0d1c62016c3949d6f7346fe561790ee47cbd2b70695cea07476938ff0" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.552011 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-api" containerID="cri-o://148c54546a54c328b2db6ab9bedc154ff074db54ed387f76af7ff0b53e27848e" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.558118 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.558449 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-log" containerID="cri-o://4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.559352 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-metadata" containerID="cri-o://c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.566029 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.575826 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jvxdd"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.579510 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jvxdd"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.586493 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-fa08-account-create-5857z"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.593446 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-4fc5f"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.598013 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-fa08-account-create-5857z"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.603273 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0627c-account-delete-z6g9h"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.609112 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-4fc5f"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.619031 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-59f9c4fc59-kvs2g"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.619295 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker-log" containerID="cri-o://f651cc8c1eb5d6258b63b8e4bc2df041f992fce6e5d8e0d734458bee6b085f54" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.619807 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker" containerID="cri-o://f1481a876ddf5f2a5996c05b5e44aa7cfa1cf846a5a9246337a1abce43f43d18" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.626397 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-627c-account-create-grrp9"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.636704 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5b4bcc558b-zx56q"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.637099 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener-log" containerID="cri-o://585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.637580 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener" containerID="cri-o://833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.641864 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-627c-account-create-grrp9"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.643863 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6547ddf7cb-m6hp6"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.644806 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6547ddf7cb-m6hp6" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api-log" containerID="cri-o://d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.647867 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6547ddf7cb-m6hp6" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api" containerID="cri-o://378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.651169 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.651399 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9895842c-0b86-4c8c-9076-3777d45634dd" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f995d871a3f41fdb1cb9c3fea81de2923a198870ea13f734a17d0294e8d0433f" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.671251 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.688898 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.724130 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder95b0-account-delete-l4qbw"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.773565 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.773655 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.773736 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6f5fbdd-0691-4ce8-888d-be0147974607" (UID: "d6f5fbdd-0691-4ce8-888d-be0147974607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.774234 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="73271be2-0048-432e-b389-183258582254" containerName="nova-scheduler-scheduler" containerID="cri-o://4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.787325 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-sb\") pod \"5091c146-f092-4878-95a3-49cb91760d17\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.787379 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-svc\") pod \"5091c146-f092-4878-95a3-49cb91760d17\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.787563 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-swift-storage-0\") pod \"5091c146-f092-4878-95a3-49cb91760d17\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.787815 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm77d\" (UniqueName: \"kubernetes.io/projected/5091c146-f092-4878-95a3-49cb91760d17-kube-api-access-dm77d\") pod \"5091c146-f092-4878-95a3-49cb91760d17\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.787899 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-config\") pod \"5091c146-f092-4878-95a3-49cb91760d17\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.787938 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-nb\") pod \"5091c146-f092-4878-95a3-49cb91760d17\" (UID: \"5091c146-f092-4878-95a3-49cb91760d17\") " Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.788404 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.788422 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.788431 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.789936 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.797673 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-77482"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.810488 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.811305 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="ebbf3e4c-75db-4d41-8220-4b5c687d815f" containerName="nova-cell0-conductor-conductor" containerID="cri-o://0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe" gracePeriod=30 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.814502 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="9895842c-0b86-4c8c-9076-3777d45634dd" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.198:6080/vnc_lite.html\": dial tcp 10.217.0.198:6080: connect: connection refused" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.828360 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-77482"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.843569 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5091c146-f092-4878-95a3-49cb91760d17-kube-api-access-dm77d" (OuterVolumeSpecName: "kube-api-access-dm77d") pod "5091c146-f092-4878-95a3-49cb91760d17" (UID: "5091c146-f092-4878-95a3-49cb91760d17"). InnerVolumeSpecName "kube-api-access-dm77d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.843774 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.896520 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.896559 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm77d\" (UniqueName: \"kubernetes.io/projected/5091c146-f092-4878-95a3-49cb91760d17-kube-api-access-dm77d\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.896572 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.919219 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ncgvz"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.966860 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="4deb0612-547f-4067-b95b-5794663d21aa" containerName="rabbitmq" containerID="cri-o://2d507eb2a2ecc2839157b138e7210f5bbddda79641955e41f82171d217de786c" gracePeriod=604800 Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.975244 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 06:52:37 crc kubenswrapper[4810]: I1008 06:52:37.975888 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" containerName="nova-cell1-conductor-conductor" containerID="cri-o://f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" gracePeriod=30 Oct 08 06:52:38 crc kubenswrapper[4810]: W1008 06:52:37.995353 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7b1ed43_571a_4b70_9eae_c19a7675bc59.slice/crio-4bb120ff7e4e8942dd8b11b3aa307906a8f1066708521c268a83b6097b787a8c WatchSource:0}: Error finding container 4bb120ff7e4e8942dd8b11b3aa307906a8f1066708521c268a83b6097b787a8c: Status 404 returned error can't find the container with id 4bb120ff7e4e8942dd8b11b3aa307906a8f1066708521c268a83b6097b787a8c Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.009116 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ncgvz"] Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.015228 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb73c-account-delete-7cpkv"] Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.029064 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi29ac-account-delete-mjvf2"] Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.032321 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.032659 4810 scope.go:117] "RemoveContainer" containerID="9285a3fd9c733d4b1075005ea98d5446430d36e9ccc354313118ef7f01c677b6" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.054024 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0627c-account-delete-z6g9h"] Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.102630 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config-secret\") pod \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.102723 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxbrr\" (UniqueName: \"kubernetes.io/projected/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-kube-api-access-fxbrr\") pod \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.102787 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config\") pod \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.102849 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-combined-ca-bundle\") pod \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\" (UID: \"31bfa5d9-dc8e-495d-983a-a17bd5a9b11f\") " Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.129929 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-kube-api-access-fxbrr" (OuterVolumeSpecName: "kube-api-access-fxbrr") pod "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" (UID: "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f"). InnerVolumeSpecName "kube-api-access-fxbrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.135251 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxbrr\" (UniqueName: \"kubernetes.io/projected/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-kube-api-access-fxbrr\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.197762 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0805f157-1607-41e2-9069-4e07dc59a46b" path="/var/lib/kubelet/pods/0805f157-1607-41e2-9069-4e07dc59a46b/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.198429 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0babfa70-91ca-45ad-afb6-aa40d2300144" path="/var/lib/kubelet/pods/0babfa70-91ca-45ad-afb6-aa40d2300144/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.199067 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e" path="/var/lib/kubelet/pods/1b5f8bd4-7e61-4cf9-9e5b-544b82b4e33e/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.199785 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d65fbac-67f5-445f-9fd6-56ae7828fba8" path="/var/lib/kubelet/pods/1d65fbac-67f5-445f-9fd6-56ae7828fba8/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.206434 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40eae6ac-32d1-406c-b971-04621dcfaeb3" path="/var/lib/kubelet/pods/40eae6ac-32d1-406c-b971-04621dcfaeb3/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.208566 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d40d07a-2b98-458f-87d9-de8db4233e85" path="/var/lib/kubelet/pods/4d40d07a-2b98-458f-87d9-de8db4233e85/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.209095 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508940e7-94fa-4ada-a7b6-0082a5fa6091" path="/var/lib/kubelet/pods/508940e7-94fa-4ada-a7b6-0082a5fa6091/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.209628 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="622e00b4-7ccf-4f15-b722-bf67388fdb83" path="/var/lib/kubelet/pods/622e00b4-7ccf-4f15-b722-bf67388fdb83/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.214952 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fe43e84-be58-40c1-8e5e-64547f990174" path="/var/lib/kubelet/pods/6fe43e84-be58-40c1-8e5e-64547f990174/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.215812 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7be74aea-5896-4b32-8f70-363522e5e7ee" path="/var/lib/kubelet/pods/7be74aea-5896-4b32-8f70-363522e5e7ee/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.216596 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8076563f-a890-4f24-8b45-c4141d27fa73" path="/var/lib/kubelet/pods/8076563f-a890-4f24-8b45-c4141d27fa73/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.217116 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29b98cd-a057-457e-91f5-aa6e7a05b840" path="/var/lib/kubelet/pods/a29b98cd-a057-457e-91f5-aa6e7a05b840/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.221024 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f82549-5e55-491b-9c01-bca9f29e25f3" path="/var/lib/kubelet/pods/a5f82549-5e55-491b-9c01-bca9f29e25f3/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.222948 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6363724-0c41-4b54-89e6-c672b91bbe98" path="/var/lib/kubelet/pods/a6363724-0c41-4b54-89e6-c672b91bbe98/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.223468 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc50ffeb-5f89-4a44-a95f-2321d36f6fb7" path="/var/lib/kubelet/pods/bc50ffeb-5f89-4a44-a95f-2321d36f6fb7/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.224373 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d12c373e-80ef-4de3-bb28-1044cfa4b8d6" path="/var/lib/kubelet/pods/d12c373e-80ef-4de3-bb28-1044cfa4b8d6/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.225533 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d318fcf9-d106-40cd-8ff9-d7e5abf43374" path="/var/lib/kubelet/pods/d318fcf9-d106-40cd-8ff9-d7e5abf43374/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.226042 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6755119-3241-4b88-b417-a13caf382092" path="/var/lib/kubelet/pods/d6755119-3241-4b88-b417-a13caf382092/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.226647 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec9c0204-fb65-420a-8917-7fc47453599f" path="/var/lib/kubelet/pods/ec9c0204-fb65-420a-8917-7fc47453599f/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.227764 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7375a67-1872-4a18-a8b6-04cc8a6f06b8" path="/var/lib/kubelet/pods/f7375a67-1872-4a18-a8b6-04cc8a6f06b8/volumes" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.326937 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="a940c7a1-b3cc-4168-99e8-750c87845314" containerName="galera" containerID="cri-o://dd78d50a7f1c3827da6046339729794dc53b8121adbd9c479e00e2f78cbf711a" gracePeriod=30 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.336725 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.349424 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.388231 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" (UID: "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.417773 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-config" (OuterVolumeSpecName: "config") pod "5091c146-f092-4878-95a3-49cb91760d17" (UID: "5091c146-f092-4878-95a3-49cb91760d17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.438712 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "d6f5fbdd-0691-4ce8-888d-be0147974607" (UID: "d6f5fbdd-0691-4ce8-888d-be0147974607"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.451395 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.451422 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f5fbdd-0691-4ce8-888d-be0147974607-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.451432 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.503811 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" (UID: "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.506871 4810 generic.go:334] "Generic (PLEG): container finished" podID="12678e95-0510-4cd6-86c8-701ea97a247e" containerID="81749551b081df35eaa781dfa19fa8b4dcb1258c7b85162b7bf3626674f4ace6" exitCode=0 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.506904 4810 generic.go:334] "Generic (PLEG): container finished" podID="12678e95-0510-4cd6-86c8-701ea97a247e" containerID="f6b4a8db2b3498b1430bdc164af3693743a79260b3c2097cdb4bee62cab40e39" exitCode=0 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.531848 4810 generic.go:334] "Generic (PLEG): container finished" podID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerID="585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f" exitCode=143 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.536262 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "a9ba387c-c5ce-4f6e-8323-7e6afce894ba" (UID: "a9ba387c-c5ce-4f6e-8323-7e6afce894ba"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.539618 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" (UID: "31bfa5d9-dc8e-495d-983a-a17bd5a9b11f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.541696 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.542173 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" (UID: "4a3bceca-c9a2-45e1-9f81-2bd9f6985c36"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.550907 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5091c146-f092-4878-95a3-49cb91760d17" (UID: "5091c146-f092-4878-95a3-49cb91760d17"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.554204 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.554244 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.554253 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ba387c-c5ce-4f6e-8323-7e6afce894ba-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.554265 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.554281 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.554290 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.569450 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.575308 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5091c146-f092-4878-95a3-49cb91760d17" (UID: "5091c146-f092-4878-95a3-49cb91760d17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.647550 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="d8f77941e09530d2318f1aa89644df82f0915aa3cfe954e93ebcfc8a0be63897" exitCode=0 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.647577 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="ce160cdae1ce3f85d7b348919fd8ec1e63ee09bafdd24ff6e4c23db7ad5a2ff1" exitCode=0 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.649399 4810 generic.go:334] "Generic (PLEG): container finished" podID="ced62b16-201e-4f75-a970-20bad14e1441" containerID="813c14b28553f034c580a431b8bf0d538af3dcb23cca601feb8f3060311ea67a" exitCode=0 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.658229 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: E1008 06:52:38.658399 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 06:52:38 crc kubenswrapper[4810]: E1008 06:52:38.658568 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data podName:2a44c96c-b96c-409c-9c9e-c049d9fe68b5 nodeName:}" failed. No retries permitted until 2025-10-08 06:52:42.658505869 +0000 UTC m=+1265.292945609 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data") pod "rabbitmq-server-0" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5") : configmap "rabbitmq-config-data" not found Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.678524 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5091c146-f092-4878-95a3-49cb91760d17" (UID: "5091c146-f092-4878-95a3-49cb91760d17"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.696584 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5091c146-f092-4878-95a3-49cb91760d17" (UID: "5091c146-f092-4878-95a3-49cb91760d17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.710434 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4a3bceca-c9a2-45e1-9f81-2bd9f6985c36/ovsdbserver-nb/0.log" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.710622 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.720099 4810 generic.go:334] "Generic (PLEG): container finished" podID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerID="4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84" exitCode=143 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.723499 4810 generic.go:334] "Generic (PLEG): container finished" podID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerID="f651cc8c1eb5d6258b63b8e4bc2df041f992fce6e5d8e0d734458bee6b085f54" exitCode=143 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.725341 4810 generic.go:334] "Generic (PLEG): container finished" podID="9895842c-0b86-4c8c-9076-3777d45634dd" containerID="f995d871a3f41fdb1cb9c3fea81de2923a198870ea13f734a17d0294e8d0433f" exitCode=0 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.726981 4810 generic.go:334] "Generic (PLEG): container finished" podID="016c35ce-3072-485f-86b8-bbe927f78f81" containerID="d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b" exitCode=143 Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.753457 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.759639 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.759680 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5091c146-f092-4878-95a3-49cb91760d17-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:38 crc kubenswrapper[4810]: E1008 06:52:38.759740 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:38 crc kubenswrapper[4810]: E1008 06:52:38.759777 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data podName:4deb0612-547f-4067-b95b-5794663d21aa nodeName:}" failed. No retries permitted until 2025-10-08 06:52:42.759764618 +0000 UTC m=+1265.394204348 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data") pod "rabbitmq-cell1-server-0" (UID: "4deb0612-547f-4067-b95b-5794663d21aa") : configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.765295 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 06:52:38 crc kubenswrapper[4810]: I1008 06:52:38.828895 4810 generic.go:334] "Generic (PLEG): container finished" podID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerID="a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df" exitCode=0 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:38.841763 4810 generic.go:334] "Generic (PLEG): container finished" podID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerID="744e6bf0d1c62016c3949d6f7346fe561790ee47cbd2b70695cea07476938ff0" exitCode=143 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.148633 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="4deb0612-547f-4067-b95b-5794663d21aa" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.238702 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.243093 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.245562 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.245598 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" containerName="nova-cell1-conductor-conductor" Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.613322 4810 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.536s" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613436 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"12678e95-0510-4cd6-86c8-701ea97a247e","Type":"ContainerDied","Data":"81749551b081df35eaa781dfa19fa8b4dcb1258c7b85162b7bf3626674f4ace6"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613479 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"12678e95-0510-4cd6-86c8-701ea97a247e","Type":"ContainerDied","Data":"f6b4a8db2b3498b1430bdc164af3693743a79260b3c2097cdb4bee62cab40e39"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613497 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7655f5c6df-qtz48"] Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613522 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" event={"ID":"aabc3b24-1971-4326-aff0-12ddbf1f5ea1","Type":"ContainerDied","Data":"585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613541 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"d8f77941e09530d2318f1aa89644df82f0915aa3cfe954e93ebcfc8a0be63897"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613557 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"ce160cdae1ce3f85d7b348919fd8ec1e63ee09bafdd24ff6e4c23db7ad5a2ff1"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc68e-account-delete-m9f52" event={"ID":"ced62b16-201e-4f75-a970-20bad14e1441","Type":"ContainerDied","Data":"813c14b28553f034c580a431b8bf0d538af3dcb23cca601feb8f3060311ea67a"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613590 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a3bceca-c9a2-45e1-9f81-2bd9f6985c36","Type":"ContainerDied","Data":"5919a28b07d4e5a1322abe30f98f1aba9bdeddb7cc6e3637d5ac8cb63ef9752d"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613607 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8832ab0b-b264-4db9-8b70-87920c1e4826","Type":"ContainerDied","Data":"4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613622 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" event={"ID":"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1","Type":"ContainerDied","Data":"f651cc8c1eb5d6258b63b8e4bc2df041f992fce6e5d8e0d734458bee6b085f54"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613637 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9895842c-0b86-4c8c-9076-3777d45634dd","Type":"ContainerDied","Data":"f995d871a3f41fdb1cb9c3fea81de2923a198870ea13f734a17d0294e8d0433f"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613652 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6547ddf7cb-m6hp6" event={"ID":"016c35ce-3072-485f-86b8-bbe927f78f81","Type":"ContainerDied","Data":"d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder95b0-account-delete-l4qbw" event={"ID":"90da3b83-b1b7-4d11-9f4d-37cf76b93edb","Type":"ContainerStarted","Data":"f1a4cf05a0397eeca7071d648a9bc8e63df17b35019fb39d5e0f9f897caf8b25"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder95b0-account-delete-l4qbw" event={"ID":"90da3b83-b1b7-4d11-9f4d-37cf76b93edb","Type":"ContainerStarted","Data":"ce86e03780d3d25ad661812da1c779f83d42f44b470b13d037b1a2dfec12b95b"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613692 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4d96bb9-89r5m" event={"ID":"5091c146-f092-4878-95a3-49cb91760d17","Type":"ContainerDied","Data":"4c5faeb23f4543d30881f9452b47e01d9f2e9be61c6a1a433d8bd4b1d0f9d199"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613705 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi29ac-account-delete-mjvf2" event={"ID":"b7b1ed43-571a-4b70-9eae-c19a7675bc59","Type":"ContainerStarted","Data":"4bb120ff7e4e8942dd8b11b3aa307906a8f1066708521c268a83b6097b787a8c"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613723 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a9ba387c-c5ce-4f6e-8323-7e6afce894ba","Type":"ContainerDied","Data":"916d488c5c8f7a36a08582b805ce1d1957f0e604d233e3dca47e1b5668fc8256"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613739 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb73c-account-delete-7cpkv" event={"ID":"9c1351c1-21aa-4004-a402-a603343c58d7","Type":"ContainerStarted","Data":"6f64faeda14fedc9020922450b7c171b79e916ba6b0c8666aae00e14e2b44da8"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613752 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0627c-account-delete-z6g9h" event={"ID":"64350cd1-38b8-4c64-8c01-105432d1b91f","Type":"ContainerStarted","Data":"bc27039f007960e307b63706ef70ac81ec2dbb5eaf21183dea8d10b1c2739a92"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613765 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f65ff9-7hhz2" event={"ID":"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0","Type":"ContainerDied","Data":"a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613782 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f266b088-0be1-4d57-b8f9-c891e8ad1ed3","Type":"ContainerDied","Data":"744e6bf0d1c62016c3949d6f7346fe561790ee47cbd2b70695cea07476938ff0"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.613807 4810 scope.go:117] "RemoveContainer" containerID="02937c8f188e832e3f815057f823449d155b5ae31271a2816bdd2a5efb1d99a8" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.628389 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7655f5c6df-qtz48" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-httpd" containerID="cri-o://67eb5ec810014b1c8f7679774ea7850e6282168a5aaf2c5cf4b56dfb7bddaec8" gracePeriod=30 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.628690 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7655f5c6df-qtz48" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-server" containerID="cri-o://497c584cc6367d012ea7e0e3c2b067d6f076bf9d675830debc76b79c7a56c2ca" gracePeriod=30 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.686360 4810 scope.go:117] "RemoveContainer" containerID="e4123f776f7b7b29b557fdc6e108f0d0848f10b0a0ef5e629afe696812883bbf" Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.771601 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe is running failed: container process not found" containerID="0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.772275 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe is running failed: container process not found" containerID="0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.772699 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe is running failed: container process not found" containerID="0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.772853 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="ebbf3e4c-75db-4d41-8220-4b5c687d815f" containerName="nova-cell0-conductor-conductor" Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.831711 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c is running failed: container process not found" containerID="4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.832228 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c is running failed: container process not found" containerID="4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.833272 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c is running failed: container process not found" containerID="4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 06:52:39 crc kubenswrapper[4810]: E1008 06:52:39.833314 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="73271be2-0048-432e-b389-183258582254" containerName="nova-scheduler-scheduler" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.865221 4810 scope.go:117] "RemoveContainer" containerID="7d7302d109e2280340776457b1d206311a58db6e8a68f517aea3acb2a3267097" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.866409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc68e-account-delete-m9f52" event={"ID":"ced62b16-201e-4f75-a970-20bad14e1441","Type":"ContainerDied","Data":"c791d0a3d7ca0fde0848db8aef544a00cf85490bb89d1e9e5985a4a3259cbdd7"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.866455 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c791d0a3d7ca0fde0848db8aef544a00cf85490bb89d1e9e5985a4a3259cbdd7" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.876312 4810 generic.go:334] "Generic (PLEG): container finished" podID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerID="f1481a876ddf5f2a5996c05b5e44aa7cfa1cf846a5a9246337a1abce43f43d18" exitCode=0 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.884403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" event={"ID":"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1","Type":"ContainerDied","Data":"f1481a876ddf5f2a5996c05b5e44aa7cfa1cf846a5a9246337a1abce43f43d18"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.884481 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" event={"ID":"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1","Type":"ContainerDied","Data":"a1da74fa58f3eb079cecd485ee70af2de2336833d2f21c9da92f00fc742fa804"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.884498 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1da74fa58f3eb079cecd485ee70af2de2336833d2f21c9da92f00fc742fa804" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.889855 4810 generic.go:334] "Generic (PLEG): container finished" podID="ebbf3e4c-75db-4d41-8220-4b5c687d815f" containerID="0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe" exitCode=0 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.890003 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebbf3e4c-75db-4d41-8220-4b5c687d815f","Type":"ContainerDied","Data":"0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.890042 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ebbf3e4c-75db-4d41-8220-4b5c687d815f","Type":"ContainerDied","Data":"0ecac1e8f1b2d9b982fb0729466a98a3325ac5db4d13bdf99ea89cd334ea098f"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.890061 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ecac1e8f1b2d9b982fb0729466a98a3325ac5db4d13bdf99ea89cd334ea098f" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.893322 4810 generic.go:334] "Generic (PLEG): container finished" podID="73271be2-0048-432e-b389-183258582254" containerID="4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c" exitCode=0 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.893413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73271be2-0048-432e-b389-183258582254","Type":"ContainerDied","Data":"4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.893440 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73271be2-0048-432e-b389-183258582254","Type":"ContainerDied","Data":"5610b5ce3416d1a08ec85bdf6223fc5f70a0e4ee6a1bcee0583bfab9b49a6967"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.893454 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5610b5ce3416d1a08ec85bdf6223fc5f70a0e4ee6a1bcee0583bfab9b49a6967" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.939889 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.950533 4810 scope.go:117] "RemoveContainer" containerID="2ae3f130783d76beb259a6e72a55a197e29b739bcfbe13913f1f50be2fef28b4" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.957013 4810 generic.go:334] "Generic (PLEG): container finished" podID="a940c7a1-b3cc-4168-99e8-750c87845314" containerID="dd78d50a7f1c3827da6046339729794dc53b8121adbd9c479e00e2f78cbf711a" exitCode=0 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.957155 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a940c7a1-b3cc-4168-99e8-750c87845314","Type":"ContainerDied","Data":"dd78d50a7f1c3827da6046339729794dc53b8121adbd9c479e00e2f78cbf711a"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.957185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a940c7a1-b3cc-4168-99e8-750c87845314","Type":"ContainerDied","Data":"79eb39754856cde25b148052d7ca0631a6f8d5b24de4ca5267650373ae53d0e3"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.957198 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79eb39754856cde25b148052d7ca0631a6f8d5b24de4ca5267650373ae53d0e3" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.960090 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9895842c-0b86-4c8c-9076-3777d45634dd","Type":"ContainerDied","Data":"74aec784c769a9420a25956a07e2b3353e806c230beaed1a19347a590ebb6112"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.960187 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementc68e-account-delete-m9f52" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.960238 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.970955 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"12678e95-0510-4cd6-86c8-701ea97a247e","Type":"ContainerDied","Data":"b2042103e24421016b0b0fcea6127c5320f9c9997792362bf446b2bf56013c9f"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.971021 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2042103e24421016b0b0fcea6127c5320f9c9997792362bf446b2bf56013c9f" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.977934 4810 generic.go:334] "Generic (PLEG): container finished" podID="90da3b83-b1b7-4d11-9f4d-37cf76b93edb" containerID="f1a4cf05a0397eeca7071d648a9bc8e63df17b35019fb39d5e0f9f897caf8b25" exitCode=0 Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.977982 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder95b0-account-delete-l4qbw" event={"ID":"90da3b83-b1b7-4d11-9f4d-37cf76b93edb","Type":"ContainerDied","Data":"f1a4cf05a0397eeca7071d648a9bc8e63df17b35019fb39d5e0f9f897caf8b25"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.978002 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder95b0-account-delete-l4qbw" event={"ID":"90da3b83-b1b7-4d11-9f4d-37cf76b93edb","Type":"ContainerDied","Data":"ce86e03780d3d25ad661812da1c779f83d42f44b470b13d037b1a2dfec12b95b"} Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.978012 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce86e03780d3d25ad661812da1c779f83d42f44b470b13d037b1a2dfec12b95b" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.984938 4810 scope.go:117] "RemoveContainer" containerID="a1c4944dcfd55a27b1f280aa67cbd8b54e526e0b2822d8a76c6d1e57a1032b88" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.988645 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.994452 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-scripts\") pod \"12678e95-0510-4cd6-86c8-701ea97a247e\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.994552 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrp8z\" (UniqueName: \"kubernetes.io/projected/ced62b16-201e-4f75-a970-20bad14e1441-kube-api-access-rrp8z\") pod \"ced62b16-201e-4f75-a970-20bad14e1441\" (UID: \"ced62b16-201e-4f75-a970-20bad14e1441\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.994637 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5t59\" (UniqueName: \"kubernetes.io/projected/9895842c-0b86-4c8c-9076-3777d45634dd-kube-api-access-r5t59\") pod \"9895842c-0b86-4c8c-9076-3777d45634dd\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.994665 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data\") pod \"12678e95-0510-4cd6-86c8-701ea97a247e\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.994797 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t692\" (UniqueName: \"kubernetes.io/projected/12678e95-0510-4cd6-86c8-701ea97a247e-kube-api-access-4t692\") pod \"12678e95-0510-4cd6-86c8-701ea97a247e\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.994861 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-config-data\") pod \"9895842c-0b86-4c8c-9076-3777d45634dd\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.994937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-nova-novncproxy-tls-certs\") pod \"9895842c-0b86-4c8c-9076-3777d45634dd\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.995000 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12678e95-0510-4cd6-86c8-701ea97a247e-etc-machine-id\") pod \"12678e95-0510-4cd6-86c8-701ea97a247e\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.995027 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data-custom\") pod \"12678e95-0510-4cd6-86c8-701ea97a247e\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.995078 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-combined-ca-bundle\") pod \"9895842c-0b86-4c8c-9076-3777d45634dd\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.995108 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-vencrypt-tls-certs\") pod \"9895842c-0b86-4c8c-9076-3777d45634dd\" (UID: \"9895842c-0b86-4c8c-9076-3777d45634dd\") " Oct 08 06:52:39 crc kubenswrapper[4810]: I1008 06:52:39.995151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-combined-ca-bundle\") pod \"12678e95-0510-4cd6-86c8-701ea97a247e\" (UID: \"12678e95-0510-4cd6-86c8-701ea97a247e\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.006027 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12678e95-0510-4cd6-86c8-701ea97a247e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "12678e95-0510-4cd6-86c8-701ea97a247e" (UID: "12678e95-0510-4cd6-86c8-701ea97a247e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.006086 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-fchxc"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.008739 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder95b0-account-delete-l4qbw" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.021913 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.048794 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.050630 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.056385 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-fchxc"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.057642 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9895842c-0b86-4c8c-9076-3777d45634dd-kube-api-access-r5t59" (OuterVolumeSpecName: "kube-api-access-r5t59") pod "9895842c-0b86-4c8c-9076-3777d45634dd" (UID: "9895842c-0b86-4c8c-9076-3777d45634dd"). InnerVolumeSpecName "kube-api-access-r5t59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.065319 4810 scope.go:117] "RemoveContainer" containerID="f995d871a3f41fdb1cb9c3fea81de2923a198870ea13f734a17d0294e8d0433f" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.065508 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.066624 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "12678e95-0510-4cd6-86c8-701ea97a247e" (UID: "12678e95-0510-4cd6-86c8-701ea97a247e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.066726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced62b16-201e-4f75-a970-20bad14e1441-kube-api-access-rrp8z" (OuterVolumeSpecName: "kube-api-access-rrp8z") pod "ced62b16-201e-4f75-a970-20bad14e1441" (UID: "ced62b16-201e-4f75-a970-20bad14e1441"). InnerVolumeSpecName "kube-api-access-rrp8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.067978 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12678e95-0510-4cd6-86c8-701ea97a247e-kube-api-access-4t692" (OuterVolumeSpecName: "kube-api-access-4t692") pod "12678e95-0510-4cd6-86c8-701ea97a247e" (UID: "12678e95-0510-4cd6-86c8-701ea97a247e"). InnerVolumeSpecName "kube-api-access-4t692". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.095536 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-scripts" (OuterVolumeSpecName: "scripts") pod "12678e95-0510-4cd6-86c8-701ea97a247e" (UID: "12678e95-0510-4cd6-86c8-701ea97a247e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.101317 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t692\" (UniqueName: \"kubernetes.io/projected/12678e95-0510-4cd6-86c8-701ea97a247e-kube-api-access-4t692\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.102097 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12678e95-0510-4cd6-86c8-701ea97a247e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.102191 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.102247 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.102309 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrp8z\" (UniqueName: \"kubernetes.io/projected/ced62b16-201e-4f75-a970-20bad14e1441-kube-api-access-rrp8z\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.102369 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5t59\" (UniqueName: \"kubernetes.io/projected/9895842c-0b86-4c8c-9076-3777d45634dd-kube-api-access-r5t59\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.133708 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31bfa5d9-dc8e-495d-983a-a17bd5a9b11f" path="/var/lib/kubelet/pods/31bfa5d9-dc8e-495d-983a-a17bd5a9b11f/volumes" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.135893 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f5fbdd-0691-4ce8-888d-be0147974607" path="/var/lib/kubelet/pods/d6f5fbdd-0691-4ce8-888d-be0147974607/volumes" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.170088 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-config-data" (OuterVolumeSpecName: "config-data") pod "9895842c-0b86-4c8c-9076-3777d45634dd" (UID: "9895842c-0b86-4c8c-9076-3777d45634dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203345 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-default\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203404 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-galera-tls-certs\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203432 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-kolla-config\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203468 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-generated\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203507 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-config-data\") pod \"73271be2-0048-432e-b389-183258582254\" (UID: \"73271be2-0048-432e-b389-183258582254\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203535 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbdqp\" (UniqueName: \"kubernetes.io/projected/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-kube-api-access-tbdqp\") pod \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203554 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data\") pod \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203589 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5nrg\" (UniqueName: \"kubernetes.io/projected/73271be2-0048-432e-b389-183258582254-kube-api-access-n5nrg\") pod \"73271be2-0048-432e-b389-183258582254\" (UID: \"73271be2-0048-432e-b389-183258582254\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203635 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfm8f\" (UniqueName: \"kubernetes.io/projected/ebbf3e4c-75db-4d41-8220-4b5c687d815f-kube-api-access-mfm8f\") pod \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203678 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-combined-ca-bundle\") pod \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203708 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nqm4\" (UniqueName: \"kubernetes.io/projected/a940c7a1-b3cc-4168-99e8-750c87845314-kube-api-access-4nqm4\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203744 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-secrets\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203772 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85rb7\" (UniqueName: \"kubernetes.io/projected/90da3b83-b1b7-4d11-9f4d-37cf76b93edb-kube-api-access-85rb7\") pod \"90da3b83-b1b7-4d11-9f4d-37cf76b93edb\" (UID: \"90da3b83-b1b7-4d11-9f4d-37cf76b93edb\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203802 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data-custom\") pod \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203822 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-logs\") pod \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203841 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203861 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-operator-scripts\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.203936 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-combined-ca-bundle\") pod \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\" (UID: \"c74ea5cd-1e87-40b3-b7f4-262fdc1779d1\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.204015 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-combined-ca-bundle\") pod \"73271be2-0048-432e-b389-183258582254\" (UID: \"73271be2-0048-432e-b389-183258582254\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.204057 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-combined-ca-bundle\") pod \"a940c7a1-b3cc-4168-99e8-750c87845314\" (UID: \"a940c7a1-b3cc-4168-99e8-750c87845314\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.204076 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-config-data\") pod \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\" (UID: \"ebbf3e4c-75db-4d41-8220-4b5c687d815f\") " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.205023 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.208458 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-logs" (OuterVolumeSpecName: "logs") pod "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" (UID: "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.234431 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.235201 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.235289 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.237329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.249294 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73271be2-0048-432e-b389-183258582254-kube-api-access-n5nrg" (OuterVolumeSpecName: "kube-api-access-n5nrg") pod "73271be2-0048-432e-b389-183258582254" (UID: "73271be2-0048-432e-b389-183258582254"). InnerVolumeSpecName "kube-api-access-n5nrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.274426 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebbf3e4c-75db-4d41-8220-4b5c687d815f-kube-api-access-mfm8f" (OuterVolumeSpecName: "kube-api-access-mfm8f") pod "ebbf3e4c-75db-4d41-8220-4b5c687d815f" (UID: "ebbf3e4c-75db-4d41-8220-4b5c687d815f"). InnerVolumeSpecName "kube-api-access-mfm8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.289360 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a940c7a1-b3cc-4168-99e8-750c87845314-kube-api-access-4nqm4" (OuterVolumeSpecName: "kube-api-access-4nqm4") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "kube-api-access-4nqm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.289754 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" (UID: "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.289750 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-secrets" (OuterVolumeSpecName: "secrets") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.290035 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90da3b83-b1b7-4d11-9f4d-37cf76b93edb-kube-api-access-85rb7" (OuterVolumeSpecName: "kube-api-access-85rb7") pod "90da3b83-b1b7-4d11-9f4d-37cf76b93edb" (UID: "90da3b83-b1b7-4d11-9f4d-37cf76b93edb"). InnerVolumeSpecName "kube-api-access-85rb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.309167 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9895842c-0b86-4c8c-9076-3777d45634dd" (UID: "9895842c-0b86-4c8c-9076-3777d45634dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312024 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312144 4810 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312201 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a940c7a1-b3cc-4168-99e8-750c87845314-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312278 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5nrg\" (UniqueName: \"kubernetes.io/projected/73271be2-0048-432e-b389-183258582254-kube-api-access-n5nrg\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312339 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfm8f\" (UniqueName: \"kubernetes.io/projected/ebbf3e4c-75db-4d41-8220-4b5c687d815f-kube-api-access-mfm8f\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312405 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nqm4\" (UniqueName: \"kubernetes.io/projected/a940c7a1-b3cc-4168-99e8-750c87845314-kube-api-access-4nqm4\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312465 4810 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312518 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85rb7\" (UniqueName: \"kubernetes.io/projected/90da3b83-b1b7-4d11-9f4d-37cf76b93edb-kube-api-access-85rb7\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312568 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312621 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.312744 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a940c7a1-b3cc-4168-99e8-750c87845314-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.329875 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.330828 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-kube-api-access-tbdqp" (OuterVolumeSpecName: "kube-api-access-tbdqp") pod "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" (UID: "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1"). InnerVolumeSpecName "kube-api-access-tbdqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.357130 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-config-data" (OuterVolumeSpecName: "config-data") pod "73271be2-0048-432e-b389-183258582254" (UID: "73271be2-0048-432e-b389-183258582254"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.414821 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.414860 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.414873 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbdqp\" (UniqueName: \"kubernetes.io/projected/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-kube-api-access-tbdqp\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.414907 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.427712 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "9895842c-0b86-4c8c-9076-3777d45634dd" (UID: "9895842c-0b86-4c8c-9076-3777d45634dd"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.491062 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebbf3e4c-75db-4d41-8220-4b5c687d815f" (UID: "ebbf3e4c-75db-4d41-8220-4b5c687d815f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.491178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-config-data" (OuterVolumeSpecName: "config-data") pod "ebbf3e4c-75db-4d41-8220-4b5c687d815f" (UID: "ebbf3e4c-75db-4d41-8220-4b5c687d815f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.499242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" (UID: "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.509576 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.517145 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "9895842c-0b86-4c8c-9076-3777d45634dd" (UID: "9895842c-0b86-4c8c-9076-3777d45634dd"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.519902 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.519934 4810 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.519945 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.519990 4810 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9895842c-0b86-4c8c-9076-3777d45634dd-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.520000 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.520009 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebbf3e4c-75db-4d41-8220-4b5c687d815f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.560941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73271be2-0048-432e-b389-183258582254" (UID: "73271be2-0048-432e-b389-183258582254"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.569543 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data" (OuterVolumeSpecName: "config-data") pod "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" (UID: "c74ea5cd-1e87-40b3-b7f4-262fdc1779d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.589136 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.624929 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.624958 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73271be2-0048-432e-b389-183258582254-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.624982 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.627140 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12678e95-0510-4cd6-86c8-701ea97a247e" (UID: "12678e95-0510-4cd6-86c8-701ea97a247e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.658100 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "a940c7a1-b3cc-4168-99e8-750c87845314" (UID: "a940c7a1-b3cc-4168-99e8-750c87845314"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.658930 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data" (OuterVolumeSpecName: "config-data") pod "12678e95-0510-4cd6-86c8-701ea97a247e" (UID: "12678e95-0510-4cd6-86c8-701ea97a247e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.734306 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.734343 4810 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a940c7a1-b3cc-4168-99e8-750c87845314-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.734353 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12678e95-0510-4cd6-86c8-701ea97a247e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818063 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d4d96bb9-89r5m"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818107 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d4d96bb9-89r5m"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818124 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818138 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818151 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818164 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818194 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4cvst"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818206 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4cvst"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818214 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi29ac-account-delete-mjvf2"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818227 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-29ac-account-create-ngrh8"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818237 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-29ac-account-create-ngrh8"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818246 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818257 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.818423 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" containerName="kube-state-metrics" containerID="cri-o://b767fec1d2cd5bc586cb8fb7ea2e613b95c815f5d05473a7605477f1e81ed54a" gracePeriod=30 Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.823602 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-central-agent" containerID="cri-o://1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495" gracePeriod=30 Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.823775 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="proxy-httpd" containerID="cri-o://9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc" gracePeriod=30 Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.823823 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="sg-core" containerID="cri-o://9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049" gracePeriod=30 Oct 08 06:52:40 crc kubenswrapper[4810]: I1008 06:52:40.823869 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-notification-agent" containerID="cri-o://6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c" gracePeriod=30 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:40.944832 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:40.945071 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="49cb68f6-1a48-4d79-90cf-34897b19dfaa" containerName="memcached" containerID="cri-o://9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3" gracePeriod=30 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:40.989175 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6547ddf7cb-m6hp6" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:40778->10.217.0.162:9311: read: connection reset by peer" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:40.989494 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6547ddf7cb-m6hp6" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:40776->10.217.0.162:9311: read: connection reset by peer" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.012405 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-5qq97"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.019018 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-5qq97"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.043388 4810 generic.go:334] "Generic (PLEG): container finished" podID="64350cd1-38b8-4c64-8c01-105432d1b91f" containerID="9bcba15966ec8198ea546c2b1b89861572f04a1d1dfbf86e9348a932b43ea9f1" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.043447 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0627c-account-delete-z6g9h" event={"ID":"64350cd1-38b8-4c64-8c01-105432d1b91f","Type":"ContainerDied","Data":"9bcba15966ec8198ea546c2b1b89861572f04a1d1dfbf86e9348a932b43ea9f1"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.046241 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-m2q9f"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.064691 4810 generic.go:334] "Generic (PLEG): container finished" podID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerID="ed04a8db57a733ab9675e113837de747b29cbd4384df9603ebc90167bf592eca" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.064771 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48ca6de6-470e-4226-b67b-e58bf64415e3","Type":"ContainerDied","Data":"ed04a8db57a733ab9675e113837de747b29cbd4384df9603ebc90167bf592eca"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.064810 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48ca6de6-470e-4226-b67b-e58bf64415e3","Type":"ContainerDied","Data":"b0c365df855d6b728d0caeb86ad0f03ae13282564f65afa5a26564e04b85aa9a"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.064826 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0c365df855d6b728d0caeb86ad0f03ae13282564f65afa5a26564e04b85aa9a" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.071159 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-m2q9f"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.071325 4810 generic.go:334] "Generic (PLEG): container finished" podID="30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" containerID="b767fec1d2cd5bc586cb8fb7ea2e613b95c815f5d05473a7605477f1e81ed54a" exitCode=2 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.071402 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce","Type":"ContainerDied","Data":"b767fec1d2cd5bc586cb8fb7ea2e613b95c815f5d05473a7605477f1e81ed54a"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.078422 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6b4dd4748f-7h7bc"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.078655 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-6b4dd4748f-7h7bc" podUID="493810ad-f07f-4871-8063-5d54cd5786ea" containerName="keystone-api" containerID="cri-o://5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e" gracePeriod=30 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.085375 4810 generic.go:334] "Generic (PLEG): container finished" podID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerID="b526413b9a8c8e6cae317ae5e79f9dced75f63320d48b38b7a09dab4a2bdcb30" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.085473 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684c4d4db6-dr9wq" event={"ID":"d7486bd0-d5fd-4d4f-8596-705cc22237d5","Type":"ContainerDied","Data":"b526413b9a8c8e6cae317ae5e79f9dced75f63320d48b38b7a09dab4a2bdcb30"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.085499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-684c4d4db6-dr9wq" event={"ID":"d7486bd0-d5fd-4d4f-8596-705cc22237d5","Type":"ContainerDied","Data":"cb36ae6379de91ab949ca3be579c93e33e60abbc00d6023c61489cd41a69f8bf"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.085511 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb36ae6379de91ab949ca3be579c93e33e60abbc00d6023c61489cd41a69f8bf" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.087881 4810 generic.go:334] "Generic (PLEG): container finished" podID="9c1351c1-21aa-4004-a402-a603343c58d7" containerID="6c82a6176a2ba5de48dbe7e2f03d6373712cc44e61a408561c87d04f36374341" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.087927 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb73c-account-delete-7cpkv" event={"ID":"9c1351c1-21aa-4004-a402-a603343c58d7","Type":"ContainerDied","Data":"6c82a6176a2ba5de48dbe7e2f03d6373712cc44e61a408561c87d04f36374341"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.126626 4810 generic.go:334] "Generic (PLEG): container finished" podID="35e33182-d302-417e-8e4e-fec4808047a4" containerID="8d3bf6f25cf1c45bf48dbb9c1a31f7c04f185f3159a9935da5533577297101f4" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.126916 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"35e33182-d302-417e-8e4e-fec4808047a4","Type":"ContainerDied","Data":"8d3bf6f25cf1c45bf48dbb9c1a31f7c04f185f3159a9935da5533577297101f4"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.127107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"35e33182-d302-417e-8e4e-fec4808047a4","Type":"ContainerDied","Data":"01df55a5848e63b58ce0cc7ad622be8c16ce171c68fd9e1c1842b12ddf8194e4"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.127122 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01df55a5848e63b58ce0cc7ad622be8c16ce171c68fd9e1c1842b12ddf8194e4" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.131496 4810 generic.go:334] "Generic (PLEG): container finished" podID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerID="497c584cc6367d012ea7e0e3c2b067d6f076bf9d675830debc76b79c7a56c2ca" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.131548 4810 generic.go:334] "Generic (PLEG): container finished" podID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerID="67eb5ec810014b1c8f7679774ea7850e6282168a5aaf2c5cf4b56dfb7bddaec8" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.131664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7655f5c6df-qtz48" event={"ID":"06237db0-424d-4bd7-ae4f-3f7af9cc4f92","Type":"ContainerDied","Data":"497c584cc6367d012ea7e0e3c2b067d6f076bf9d675830debc76b79c7a56c2ca"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.133928 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7655f5c6df-qtz48" event={"ID":"06237db0-424d-4bd7-ae4f-3f7af9cc4f92","Type":"ContainerDied","Data":"67eb5ec810014b1c8f7679774ea7850e6282168a5aaf2c5cf4b56dfb7bddaec8"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.133981 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi29ac-account-delete-mjvf2" event={"ID":"b7b1ed43-571a-4b70-9eae-c19a7675bc59","Type":"ContainerDied","Data":"d7bf8db66c158fdcb8ae6f04072efabe1937b00551aceb3c344d7a04e3d408fd"} Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.133850 4810 generic.go:334] "Generic (PLEG): container finished" podID="b7b1ed43-571a-4b70-9eae-c19a7675bc59" containerID="d7bf8db66c158fdcb8ae6f04072efabe1937b00551aceb3c344d7a04e3d408fd" exitCode=0 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.143474 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.144236 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementc68e-account-delete-m9f52" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.145406 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.145955 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.146040 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.150589 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder95b0-account-delete-l4qbw" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.150739 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-59f9c4fc59-kvs2g" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.159158 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.225098 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-kfjh7"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.235501 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-kfjh7"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.251596 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-eac1-account-create-bqdsm"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.262035 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-eac1-account-create-bqdsm"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.446783 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerName="galera" containerID="cri-o://70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb" gracePeriod=30 Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.749404 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.765382 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.809877 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.837164 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementc68e-account-delete-m9f52"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.872329 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementc68e-account-delete-m9f52"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.886353 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.905039 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.914585 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.925648 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.949601 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-run-httpd\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963691 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvp5m\" (UniqueName: \"kubernetes.io/projected/48ca6de6-470e-4226-b67b-e58bf64415e3-kube-api-access-rvp5m\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963724 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-internal-tls-certs\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963748 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-config-data\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963769 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963798 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-combined-ca-bundle\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963818 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data-custom\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963857 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-combined-ca-bundle\") pod \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963880 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35e33182-d302-417e-8e4e-fec4808047a4-etc-machine-id\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963894 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7486bd0-d5fd-4d4f-8596-705cc22237d5-logs\") pod \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963921 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-combined-ca-bundle\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.963935 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-public-tls-certs\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965401 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnctt\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-kube-api-access-xnctt\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965446 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-combined-ca-bundle\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965474 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-logs\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965500 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-config-data\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965516 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-scripts\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965539 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e33182-d302-417e-8e4e-fec4808047a4-logs\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965563 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-public-tls-certs\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965583 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-config-data\") pod \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965607 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-internal-tls-certs\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965627 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-scripts\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965654 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzvtm\" (UniqueName: \"kubernetes.io/projected/35e33182-d302-417e-8e4e-fec4808047a4-kube-api-access-vzvtm\") pod \"35e33182-d302-417e-8e4e-fec4808047a4\" (UID: \"35e33182-d302-417e-8e4e-fec4808047a4\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965670 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-internal-tls-certs\") pod \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965689 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-httpd-run\") pod \"48ca6de6-470e-4226-b67b-e58bf64415e3\" (UID: \"48ca6de6-470e-4226-b67b-e58bf64415e3\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965706 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-public-tls-certs\") pod \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965763 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxg6f\" (UniqueName: \"kubernetes.io/projected/d7486bd0-d5fd-4d4f-8596-705cc22237d5-kube-api-access-qxg6f\") pod \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-etc-swift\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965794 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-internal-tls-certs\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965832 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-log-httpd\") pod \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\" (UID: \"06237db0-424d-4bd7-ae4f-3f7af9cc4f92\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.965861 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-scripts\") pod \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\" (UID: \"d7486bd0-d5fd-4d4f-8596-705cc22237d5\") " Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.971813 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-59f9c4fc59-kvs2g"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.972401 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-logs" (OuterVolumeSpecName: "logs") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.972662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.973502 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e33182-d302-417e-8e4e-fec4808047a4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.976585 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7486bd0-d5fd-4d4f-8596-705cc22237d5-logs" (OuterVolumeSpecName: "logs") pod "d7486bd0-d5fd-4d4f-8596-705cc22237d5" (UID: "d7486bd0-d5fd-4d4f-8596-705cc22237d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.977056 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-scripts" (OuterVolumeSpecName: "scripts") pod "d7486bd0-d5fd-4d4f-8596-705cc22237d5" (UID: "d7486bd0-d5fd-4d4f-8596-705cc22237d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.984293 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-59f9c4fc59-kvs2g"] Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.985393 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ca6de6-470e-4226-b67b-e58bf64415e3-kube-api-access-rvp5m" (OuterVolumeSpecName: "kube-api-access-rvp5m") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "kube-api-access-rvp5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.986438 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0627c-account-delete-z6g9h" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.996355 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-kube-api-access-xnctt" (OuterVolumeSpecName: "kube-api-access-xnctt") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "kube-api-access-xnctt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.997778 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:41 crc kubenswrapper[4810]: I1008 06:52:41.998072 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35e33182-d302-417e-8e4e-fec4808047a4-logs" (OuterVolumeSpecName: "logs") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.002684 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.003877 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb73c-account-delete-7cpkv" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.017797 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-scripts" (OuterVolumeSpecName: "scripts") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.023037 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.027606 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7486bd0-d5fd-4d4f-8596-705cc22237d5-kube-api-access-qxg6f" (OuterVolumeSpecName: "kube-api-access-qxg6f") pod "d7486bd0-d5fd-4d4f-8596-705cc22237d5" (UID: "d7486bd0-d5fd-4d4f-8596-705cc22237d5"). InnerVolumeSpecName "kube-api-access-qxg6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.028063 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder95b0-account-delete-l4qbw"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.030382 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.033890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.042683 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.048550 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-scripts" (OuterVolumeSpecName: "scripts") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.049081 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e33182-d302-417e-8e4e-fec4808047a4-kube-api-access-vzvtm" (OuterVolumeSpecName: "kube-api-access-vzvtm") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "kube-api-access-vzvtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.050134 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.056636 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder95b0-account-delete-l4qbw"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.071252 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qlpt\" (UniqueName: \"kubernetes.io/projected/8832ab0b-b264-4db9-8b70-87920c1e4826-kube-api-access-4qlpt\") pod \"8832ab0b-b264-4db9-8b70-87920c1e4826\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.071381 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl2fl\" (UniqueName: \"kubernetes.io/projected/9c1351c1-21aa-4004-a402-a603343c58d7-kube-api-access-fl2fl\") pod \"9c1351c1-21aa-4004-a402-a603343c58d7\" (UID: \"9c1351c1-21aa-4004-a402-a603343c58d7\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.071444 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6942w\" (UniqueName: \"kubernetes.io/projected/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-api-access-6942w\") pod \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.071599 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8832ab0b-b264-4db9-8b70-87920c1e4826-logs\") pod \"8832ab0b-b264-4db9-8b70-87920c1e4826\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.071985 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-nova-metadata-tls-certs\") pod \"8832ab0b-b264-4db9-8b70-87920c1e4826\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.073529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-config\") pod \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.073581 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-config-data\") pod \"8832ab0b-b264-4db9-8b70-87920c1e4826\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.073637 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-combined-ca-bundle\") pod \"8832ab0b-b264-4db9-8b70-87920c1e4826\" (UID: \"8832ab0b-b264-4db9-8b70-87920c1e4826\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.073685 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-certs\") pod \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.073720 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-combined-ca-bundle\") pod \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\" (UID: \"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.073782 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrfhz\" (UniqueName: \"kubernetes.io/projected/64350cd1-38b8-4c64-8c01-105432d1b91f-kube-api-access-jrfhz\") pod \"64350cd1-38b8-4c64-8c01-105432d1b91f\" (UID: \"64350cd1-38b8-4c64-8c01-105432d1b91f\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074289 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074310 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzvtm\" (UniqueName: \"kubernetes.io/projected/35e33182-d302-417e-8e4e-fec4808047a4-kube-api-access-vzvtm\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074322 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074335 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxg6f\" (UniqueName: \"kubernetes.io/projected/d7486bd0-d5fd-4d4f-8596-705cc22237d5-kube-api-access-qxg6f\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074347 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074358 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074370 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074383 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074395 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvp5m\" (UniqueName: \"kubernetes.io/projected/48ca6de6-470e-4226-b67b-e58bf64415e3-kube-api-access-rvp5m\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074419 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074433 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074445 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/35e33182-d302-417e-8e4e-fec4808047a4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074462 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7486bd0-d5fd-4d4f-8596-705cc22237d5-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074473 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnctt\" (UniqueName: \"kubernetes.io/projected/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-kube-api-access-xnctt\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074483 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48ca6de6-470e-4226-b67b-e58bf64415e3-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074493 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.074505 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35e33182-d302-417e-8e4e-fec4808047a4-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.072305 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.085270 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8832ab0b-b264-4db9-8b70-87920c1e4826-logs" (OuterVolumeSpecName: "logs") pod "8832ab0b-b264-4db9-8b70-87920c1e4826" (UID: "8832ab0b-b264-4db9-8b70-87920c1e4826"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.085822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.086342 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c1351c1-21aa-4004-a402-a603343c58d7-kube-api-access-fl2fl" (OuterVolumeSpecName: "kube-api-access-fl2fl") pod "9c1351c1-21aa-4004-a402-a603343c58d7" (UID: "9c1351c1-21aa-4004-a402-a603343c58d7"). InnerVolumeSpecName "kube-api-access-fl2fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.100501 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-api-access-6942w" (OuterVolumeSpecName: "kube-api-access-6942w") pod "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" (UID: "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce"). InnerVolumeSpecName "kube-api-access-6942w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.110311 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64350cd1-38b8-4c64-8c01-105432d1b91f-kube-api-access-jrfhz" (OuterVolumeSpecName: "kube-api-access-jrfhz") pod "64350cd1-38b8-4c64-8c01-105432d1b91f" (UID: "64350cd1-38b8-4c64-8c01-105432d1b91f"). InnerVolumeSpecName "kube-api-access-jrfhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.110738 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ee716b-7145-415d-a330-d5f9f5c63f36" path="/var/lib/kubelet/pods/01ee716b-7145-415d-a330-d5f9f5c63f36/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.111806 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="315506dc-19f7-4299-b56f-8584a4021a70" path="/var/lib/kubelet/pods/315506dc-19f7-4299-b56f-8584a4021a70/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.112485 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" path="/var/lib/kubelet/pods/4a3bceca-c9a2-45e1-9f81-2bd9f6985c36/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.113673 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5091c146-f092-4878-95a3-49cb91760d17" path="/var/lib/kubelet/pods/5091c146-f092-4878-95a3-49cb91760d17/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.114253 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5afd09d8-444a-4cc9-a25c-fe3a0b0d8477" path="/var/lib/kubelet/pods/5afd09d8-444a-4cc9-a25c-fe3a0b0d8477/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.114704 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dce6778-383f-4471-b852-16eaed68e455" path="/var/lib/kubelet/pods/6dce6778-383f-4471-b852-16eaed68e455/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.114764 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.115366 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73271be2-0048-432e-b389-183258582254" path="/var/lib/kubelet/pods/73271be2-0048-432e-b389-183258582254/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.116269 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90da3b83-b1b7-4d11-9f4d-37cf76b93edb" path="/var/lib/kubelet/pods/90da3b83-b1b7-4d11-9f4d-37cf76b93edb/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.116752 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9895842c-0b86-4c8c-9076-3777d45634dd" path="/var/lib/kubelet/pods/9895842c-0b86-4c8c-9076-3777d45634dd/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.117402 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3b9863-3ccd-4bda-93da-ee17ac885040" path="/var/lib/kubelet/pods/9d3b9863-3ccd-4bda-93da-ee17ac885040/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.118909 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" path="/var/lib/kubelet/pods/a9ba387c-c5ce-4f6e-8323-7e6afce894ba/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.119582 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" path="/var/lib/kubelet/pods/c74ea5cd-1e87-40b3-b7f4-262fdc1779d1/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.120113 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced62b16-201e-4f75-a970-20bad14e1441" path="/var/lib/kubelet/pods/ced62b16-201e-4f75-a970-20bad14e1441/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.126740 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1f667ff-3e8b-48b9-be0d-5deb9483050a" path="/var/lib/kubelet/pods/e1f667ff-3e8b-48b9-be0d-5deb9483050a/volumes" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.140179 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8832ab0b-b264-4db9-8b70-87920c1e4826-kube-api-access-4qlpt" (OuterVolumeSpecName: "kube-api-access-4qlpt") pod "8832ab0b-b264-4db9-8b70-87920c1e4826" (UID: "8832ab0b-b264-4db9-8b70-87920c1e4826"). InnerVolumeSpecName "kube-api-access-4qlpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.163939 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7655f5c6df-qtz48" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.164666 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data" (OuterVolumeSpecName: "config-data") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.166903 4810 generic.go:334] "Generic (PLEG): container finished" podID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerID="26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e" exitCode=0 Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.167081 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.174661 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronb73c-account-delete-7cpkv" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175310 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khkkw\" (UniqueName: \"kubernetes.io/projected/dced1083-d856-4f53-bb6e-19fc64ba95a5-kube-api-access-khkkw\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175355 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016c35ce-3072-485f-86b8-bbe927f78f81-logs\") pod \"016c35ce-3072-485f-86b8-bbe927f78f81\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175376 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-internal-tls-certs\") pod \"016c35ce-3072-485f-86b8-bbe927f78f81\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175458 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-logs\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175516 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data\") pod \"016c35ce-3072-485f-86b8-bbe927f78f81\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175544 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-config-data\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175578 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-combined-ca-bundle\") pod \"016c35ce-3072-485f-86b8-bbe927f78f81\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175594 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-scripts\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175614 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-public-tls-certs\") pod \"016c35ce-3072-485f-86b8-bbe927f78f81\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175636 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data-custom\") pod \"016c35ce-3072-485f-86b8-bbe927f78f81\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175673 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwhs4\" (UniqueName: \"kubernetes.io/projected/016c35ce-3072-485f-86b8-bbe927f78f81-kube-api-access-lwhs4\") pod \"016c35ce-3072-485f-86b8-bbe927f78f81\" (UID: \"016c35ce-3072-485f-86b8-bbe927f78f81\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175724 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-httpd-run\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175761 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175793 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-combined-ca-bundle\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.175841 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-public-tls-certs\") pod \"dced1083-d856-4f53-bb6e-19fc64ba95a5\" (UID: \"dced1083-d856-4f53-bb6e-19fc64ba95a5\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.176222 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6942w\" (UniqueName: \"kubernetes.io/projected/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-api-access-6942w\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.176235 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8832ab0b-b264-4db9-8b70-87920c1e4826-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.176244 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrfhz\" (UniqueName: \"kubernetes.io/projected/64350cd1-38b8-4c64-8c01-105432d1b91f-kube-api-access-jrfhz\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.176253 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qlpt\" (UniqueName: \"kubernetes.io/projected/8832ab0b-b264-4db9-8b70-87920c1e4826-kube-api-access-4qlpt\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.176261 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.176271 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl2fl\" (UniqueName: \"kubernetes.io/projected/9c1351c1-21aa-4004-a402-a603343c58d7-kube-api-access-fl2fl\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.176280 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.179765 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/016c35ce-3072-485f-86b8-bbe927f78f81-logs" (OuterVolumeSpecName: "logs") pod "016c35ce-3072-485f-86b8-bbe927f78f81" (UID: "016c35ce-3072-485f-86b8-bbe927f78f81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.180026 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.187131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-logs" (OuterVolumeSpecName: "logs") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.188222 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dced1083-d856-4f53-bb6e-19fc64ba95a5-kube-api-access-khkkw" (OuterVolumeSpecName: "kube-api-access-khkkw") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "kube-api-access-khkkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.188772 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-scripts" (OuterVolumeSpecName: "scripts") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.192571 4810 generic.go:334] "Generic (PLEG): container finished" podID="016c35ce-3072-485f-86b8-bbe927f78f81" containerID="378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2" exitCode=0 Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.192706 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6547ddf7cb-m6hp6" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.199584 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "016c35ce-3072-485f-86b8-bbe927f78f81" (UID: "016c35ce-3072-485f-86b8-bbe927f78f81"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.203282 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/016c35ce-3072-485f-86b8-bbe927f78f81-kube-api-access-lwhs4" (OuterVolumeSpecName: "kube-api-access-lwhs4") pod "016c35ce-3072-485f-86b8-bbe927f78f81" (UID: "016c35ce-3072-485f-86b8-bbe927f78f81"). InnerVolumeSpecName "kube-api-access-lwhs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.208294 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0627c-account-delete-z6g9h" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216189 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216228 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216243 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216256 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216275 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7655f5c6df-qtz48" event={"ID":"06237db0-424d-4bd7-ae4f-3f7af9cc4f92","Type":"ContainerDied","Data":"7657cdad3606aa5cab96c05e2cbd935fffbbc42a1a0d867410bb359ed6fc9406"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216300 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216313 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dced1083-d856-4f53-bb6e-19fc64ba95a5","Type":"ContainerDied","Data":"26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216325 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dced1083-d856-4f53-bb6e-19fc64ba95a5","Type":"ContainerDied","Data":"540bad4204615250877efc93603d9fe908a64a8af2c69739d9838bbce3cab6b5"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216337 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronb73c-account-delete-7cpkv" event={"ID":"9c1351c1-21aa-4004-a402-a603343c58d7","Type":"ContainerDied","Data":"6f64faeda14fedc9020922450b7c171b79e916ba6b0c8666aae00e14e2b44da8"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216350 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6547ddf7cb-m6hp6" event={"ID":"016c35ce-3072-485f-86b8-bbe927f78f81","Type":"ContainerDied","Data":"378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216364 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6547ddf7cb-m6hp6" event={"ID":"016c35ce-3072-485f-86b8-bbe927f78f81","Type":"ContainerDied","Data":"dde73594c1882d70f3f7fe51a75d2da5070bc756ffd409defa4bcc2e723d7a85"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0627c-account-delete-z6g9h" event={"ID":"64350cd1-38b8-4c64-8c01-105432d1b91f","Type":"ContainerDied","Data":"bc27039f007960e307b63706ef70ac81ec2dbb5eaf21183dea8d10b1c2739a92"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.216397 4810 scope.go:117] "RemoveContainer" containerID="497c584cc6367d012ea7e0e3c2b067d6f076bf9d675830debc76b79c7a56c2ca" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.219106 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-config-data" (OuterVolumeSpecName: "config-data") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.224237 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.229342 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.230023 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerDied","Data":"9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.225486 4810 generic.go:334] "Generic (PLEG): container finished" podID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerID="9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc" exitCode=0 Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.233448 4810 generic.go:334] "Generic (PLEG): container finished" podID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerID="9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049" exitCode=2 Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.233458 4810 generic.go:334] "Generic (PLEG): container finished" podID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerID="1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495" exitCode=0 Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.233530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerDied","Data":"9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.233558 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerDied","Data":"1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.235634 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.235759 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.235829 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce","Type":"ContainerDied","Data":"72d60088be3097409ba2de5a4161245156bc48cdfc3bb9b38d7ae48729a5d505"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.238889 4810 generic.go:334] "Generic (PLEG): container finished" podID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerID="c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4" exitCode=0 Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.238998 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-684c4d4db6-dr9wq" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.239048 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8832ab0b-b264-4db9-8b70-87920c1e4826","Type":"ContainerDied","Data":"c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.239110 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8832ab0b-b264-4db9-8b70-87920c1e4826","Type":"ContainerDied","Data":"da193473bb6de6b14dbdc891fa8dc66336bfed450ac1381353ef6ee39f88c7c0"} Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.239713 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.240221 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.241114 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.264672 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.268647 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.268730 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.269472 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.270100 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.271168 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.271195 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.274039 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.274130 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.280207 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.280266 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="ovn-northd" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.284905 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290393 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290452 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwhs4\" (UniqueName: \"kubernetes.io/projected/016c35ce-3072-485f-86b8-bbe927f78f81-kube-api-access-lwhs4\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290469 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290480 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290497 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290526 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290537 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290556 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khkkw\" (UniqueName: \"kubernetes.io/projected/dced1083-d856-4f53-bb6e-19fc64ba95a5-kube-api-access-khkkw\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290565 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016c35ce-3072-485f-86b8-bbe927f78f81-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.290574 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dced1083-d856-4f53-bb6e-19fc64ba95a5-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.292769 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.301846 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="49cb68f6-1a48-4d79-90cf-34897b19dfaa" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.108:11211: connect: connection refused" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.292996 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.330734 4810 scope.go:117] "RemoveContainer" containerID="67eb5ec810014b1c8f7679774ea7850e6282168a5aaf2c5cf4b56dfb7bddaec8" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.330884 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronb73c-account-delete-7cpkv"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.337044 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8832ab0b-b264-4db9-8b70-87920c1e4826" (UID: "8832ab0b-b264-4db9-8b70-87920c1e4826"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.342915 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronb73c-account-delete-7cpkv"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.344929 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.375874 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0627c-account-delete-z6g9h"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.381823 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0627c-account-delete-z6g9h"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.387678 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-config-data" (OuterVolumeSpecName: "config-data") pod "8832ab0b-b264-4db9-8b70-87920c1e4826" (UID: "8832ab0b-b264-4db9-8b70-87920c1e4826"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.406159 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.406195 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.406206 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.407838 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8832ab0b-b264-4db9-8b70-87920c1e4826" (UID: "8832ab0b-b264-4db9-8b70-87920c1e4826"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.415311 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-config-data" (OuterVolumeSpecName: "config-data") pod "d7486bd0-d5fd-4d4f-8596-705cc22237d5" (UID: "d7486bd0-d5fd-4d4f-8596-705cc22237d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.421253 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.423679 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.507247 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.510721 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.510789 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.510802 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.510840 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.510855 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8832ab0b-b264-4db9-8b70-87920c1e4826-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.513010 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" (UID: "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.549991 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" (UID: "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.595250 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "016c35ce-3072-485f-86b8-bbe927f78f81" (UID: "016c35ce-3072-485f-86b8-bbe927f78f81"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.612786 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.612823 4810 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.612836 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.633258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.643629 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.654980 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.673197 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-config-data" (OuterVolumeSpecName: "config-data") pod "48ca6de6-470e-4226-b67b-e58bf64415e3" (UID: "48ca6de6-470e-4226-b67b-e58bf64415e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.691453 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "016c35ce-3072-485f-86b8-bbe927f78f81" (UID: "016c35ce-3072-485f-86b8-bbe927f78f81"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.699078 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "35e33182-d302-417e-8e4e-fec4808047a4" (UID: "35e33182-d302-417e-8e4e-fec4808047a4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.703794 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7486bd0-d5fd-4d4f-8596-705cc22237d5" (UID: "d7486bd0-d5fd-4d4f-8596-705cc22237d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714200 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "016c35ce-3072-485f-86b8-bbe927f78f81" (UID: "016c35ce-3072-485f-86b8-bbe927f78f81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714709 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48ca6de6-470e-4226-b67b-e58bf64415e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714748 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714757 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714765 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714774 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714785 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714792 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.714801 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35e33182-d302-417e-8e4e-fec4808047a4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.714885 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.714933 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data podName:2a44c96c-b96c-409c-9c9e-c049d9fe68b5 nodeName:}" failed. No retries permitted until 2025-10-08 06:52:50.714918578 +0000 UTC m=+1273.349358318 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data") pod "rabbitmq-server-0" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5") : configmap "rabbitmq-config-data" not found Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.728474 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" (UID: "30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.733251 4810 scope.go:117] "RemoveContainer" containerID="26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.745201 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06237db0-424d-4bd7-ae4f-3f7af9cc4f92" (UID: "06237db0-424d-4bd7-ae4f-3f7af9cc4f92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.754420 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.758852 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi29ac-account-delete-mjvf2" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.761135 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-qtj9x" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" probeResult="failure" output="command timed out" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.762047 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-config-data" (OuterVolumeSpecName: "config-data") pod "dced1083-d856-4f53-bb6e-19fc64ba95a5" (UID: "dced1083-d856-4f53-bb6e-19fc64ba95a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.763185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data" (OuterVolumeSpecName: "config-data") pod "016c35ce-3072-485f-86b8-bbe927f78f81" (UID: "016c35ce-3072-485f-86b8-bbe927f78f81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.790460 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d7486bd0-d5fd-4d4f-8596-705cc22237d5" (UID: "d7486bd0-d5fd-4d4f-8596-705cc22237d5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.800036 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.809082 4810 scope.go:117] "RemoveContainer" containerID="54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.809550 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.809761 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-qtj9x" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" probeResult="failure" output=< Oct 08 06:52:42 crc kubenswrapper[4810]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Oct 08 06:52:42 crc kubenswrapper[4810]: > Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.820241 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gm5l\" (UniqueName: \"kubernetes.io/projected/b7b1ed43-571a-4b70-9eae-c19a7675bc59-kube-api-access-4gm5l\") pod \"b7b1ed43-571a-4b70-9eae-c19a7675bc59\" (UID: \"b7b1ed43-571a-4b70-9eae-c19a7675bc59\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.821124 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.821215 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/016c35ce-3072-485f-86b8-bbe927f78f81-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.821269 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dced1083-d856-4f53-bb6e-19fc64ba95a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.821327 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06237db0-424d-4bd7-ae4f-3f7af9cc4f92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.821378 4810 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.821493 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.821623 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data podName:4deb0612-547f-4067-b95b-5794663d21aa nodeName:}" failed. No retries permitted until 2025-10-08 06:52:50.821606967 +0000 UTC m=+1273.456046707 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data") pod "rabbitmq-cell1-server-0" (UID: "4deb0612-547f-4067-b95b-5794663d21aa") : configmap "rabbitmq-cell1-config-data" not found Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.830808 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b1ed43-571a-4b70-9eae-c19a7675bc59-kube-api-access-4gm5l" (OuterVolumeSpecName: "kube-api-access-4gm5l") pod "b7b1ed43-571a-4b70-9eae-c19a7675bc59" (UID: "b7b1ed43-571a-4b70-9eae-c19a7675bc59"). InnerVolumeSpecName "kube-api-access-4gm5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.852495 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d7486bd0-d5fd-4d4f-8596-705cc22237d5" (UID: "d7486bd0-d5fd-4d4f-8596-705cc22237d5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.866154 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.879353 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.891172 4810 scope.go:117] "RemoveContainer" containerID="26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.891771 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e\": container with ID starting with 26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e not found: ID does not exist" containerID="26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.891809 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e"} err="failed to get container status \"26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e\": rpc error: code = NotFound desc = could not find container \"26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e\": container with ID starting with 26163e817f4c57dedba5d9c3eb9fdf00cd3e1a91d252140bc14e8d53b851c12e not found: ID does not exist" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.891827 4810 scope.go:117] "RemoveContainer" containerID="54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d" Oct 08 06:52:42 crc kubenswrapper[4810]: E1008 06:52:42.893289 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d\": container with ID starting with 54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d not found: ID does not exist" containerID="54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.893316 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d"} err="failed to get container status \"54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d\": rpc error: code = NotFound desc = could not find container \"54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d\": container with ID starting with 54d3a0c67537b87fc58d729e0898be280dc997db5e53128f7e72f60cccd3397d not found: ID does not exist" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.893331 4810 scope.go:117] "RemoveContainer" containerID="6c82a6176a2ba5de48dbe7e2f03d6373712cc44e61a408561c87d04f36374341" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.922188 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kolla-config\") pod \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.922235 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ssjm\" (UniqueName: \"kubernetes.io/projected/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kube-api-access-7ssjm\") pod \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.922266 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-combined-ca-bundle\") pod \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.922291 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-memcached-tls-certs\") pod \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.922327 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-config-data\") pod \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\" (UID: \"49cb68f6-1a48-4d79-90cf-34897b19dfaa\") " Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.922776 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gm5l\" (UniqueName: \"kubernetes.io/projected/b7b1ed43-571a-4b70-9eae-c19a7675bc59-kube-api-access-4gm5l\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.922790 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7486bd0-d5fd-4d4f-8596-705cc22237d5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.924283 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "49cb68f6-1a48-4d79-90cf-34897b19dfaa" (UID: "49cb68f6-1a48-4d79-90cf-34897b19dfaa"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.927198 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-config-data" (OuterVolumeSpecName: "config-data") pod "49cb68f6-1a48-4d79-90cf-34897b19dfaa" (UID: "49cb68f6-1a48-4d79-90cf-34897b19dfaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.932164 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kube-api-access-7ssjm" (OuterVolumeSpecName: "kube-api-access-7ssjm") pod "49cb68f6-1a48-4d79-90cf-34897b19dfaa" (UID: "49cb68f6-1a48-4d79-90cf-34897b19dfaa"). InnerVolumeSpecName "kube-api-access-7ssjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.935017 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.201:3000/\": dial tcp 10.217.0.201:3000: connect: connection refused" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.949685 4810 scope.go:117] "RemoveContainer" containerID="378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.952096 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49cb68f6-1a48-4d79-90cf-34897b19dfaa" (UID: "49cb68f6-1a48-4d79-90cf-34897b19dfaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.962935 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7655f5c6df-qtz48"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.983938 4810 scope.go:117] "RemoveContainer" containerID="d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b" Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.984668 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7655f5c6df-qtz48"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.990988 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:52:42 crc kubenswrapper[4810]: I1008 06:52:42.998186 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.002547 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "49cb68f6-1a48-4d79-90cf-34897b19dfaa" (UID: "49cb68f6-1a48-4d79-90cf-34897b19dfaa"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.006020 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6547ddf7cb-m6hp6"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.009109 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6547ddf7cb-m6hp6"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.012666 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.017069 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.024119 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.028423 4810 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.028547 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ssjm\" (UniqueName: \"kubernetes.io/projected/49cb68f6-1a48-4d79-90cf-34897b19dfaa-kube-api-access-7ssjm\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.028626 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.028701 4810 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49cb68f6-1a48-4d79-90cf-34897b19dfaa-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.028791 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49cb68f6-1a48-4d79-90cf-34897b19dfaa-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.030262 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.046655 4810 scope.go:117] "RemoveContainer" containerID="378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2" Oct 08 06:52:43 crc kubenswrapper[4810]: E1008 06:52:43.047338 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2\": container with ID starting with 378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2 not found: ID does not exist" containerID="378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.047371 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2"} err="failed to get container status \"378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2\": rpc error: code = NotFound desc = could not find container \"378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2\": container with ID starting with 378977ac9019e2ddabc10cc34b31da7a458c042b6e522b99ba8fa051b7bb8eb2 not found: ID does not exist" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.047394 4810 scope.go:117] "RemoveContainer" containerID="d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b" Oct 08 06:52:43 crc kubenswrapper[4810]: E1008 06:52:43.047780 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b\": container with ID starting with d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b not found: ID does not exist" containerID="d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.047802 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b"} err="failed to get container status \"d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b\": rpc error: code = NotFound desc = could not find container \"d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b\": container with ID starting with d2f5d40ce9a753574ef9cc18a0c484e8aa408b569063d39ec757a893ac556c3b not found: ID does not exist" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.047820 4810 scope.go:117] "RemoveContainer" containerID="9bcba15966ec8198ea546c2b1b89861572f04a1d1dfbf86e9348a932b43ea9f1" Oct 08 06:52:43 crc kubenswrapper[4810]: W1008 06:52:43.076789 4810 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64350cd1_38b8_4c64_8c01_105432d1b91f.slice/crio-conmon-9bcba15966ec8198ea546c2b1b89861572f04a1d1dfbf86e9348a932b43ea9f1.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64350cd1_38b8_4c64_8c01_105432d1b91f.slice/crio-conmon-9bcba15966ec8198ea546c2b1b89861572f04a1d1dfbf86e9348a932b43ea9f1.scope: no such file or directory Oct 08 06:52:43 crc kubenswrapper[4810]: W1008 06:52:43.076876 4810 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7b1ed43_571a_4b70_9eae_c19a7675bc59.slice/crio-conmon-d7bf8db66c158fdcb8ae6f04072efabe1937b00551aceb3c344d7a04e3d408fd.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7b1ed43_571a_4b70_9eae_c19a7675bc59.slice/crio-conmon-d7bf8db66c158fdcb8ae6f04072efabe1937b00551aceb3c344d7a04e3d408fd.scope: no such file or directory Oct 08 06:52:43 crc kubenswrapper[4810]: W1008 06:52:43.077335 4810 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1351c1_21aa_4004_a402_a603343c58d7.slice/crio-6c82a6176a2ba5de48dbe7e2f03d6373712cc44e61a408561c87d04f36374341.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1351c1_21aa_4004_a402_a603343c58d7.slice/crio-6c82a6176a2ba5de48dbe7e2f03d6373712cc44e61a408561c87d04f36374341.scope: no such file or directory Oct 08 06:52:43 crc kubenswrapper[4810]: W1008 06:52:43.077365 4810 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64350cd1_38b8_4c64_8c01_105432d1b91f.slice/crio-9bcba15966ec8198ea546c2b1b89861572f04a1d1dfbf86e9348a932b43ea9f1.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64350cd1_38b8_4c64_8c01_105432d1b91f.slice/crio-9bcba15966ec8198ea546c2b1b89861572f04a1d1dfbf86e9348a932b43ea9f1.scope: no such file or directory Oct 08 06:52:43 crc kubenswrapper[4810]: W1008 06:52:43.077382 4810 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7b1ed43_571a_4b70_9eae_c19a7675bc59.slice/crio-d7bf8db66c158fdcb8ae6f04072efabe1937b00551aceb3c344d7a04e3d408fd.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7b1ed43_571a_4b70_9eae_c19a7675bc59.slice/crio-d7bf8db66c158fdcb8ae6f04072efabe1937b00551aceb3c344d7a04e3d408fd.scope: no such file or directory Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.087427 4810 scope.go:117] "RemoveContainer" containerID="b767fec1d2cd5bc586cb8fb7ea2e613b95c815f5d05473a7605477f1e81ed54a" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.133271 4810 scope.go:117] "RemoveContainer" containerID="c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.170738 4810 scope.go:117] "RemoveContainer" containerID="4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.194465 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-684c4d4db6-dr9wq"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.198931 4810 scope.go:117] "RemoveContainer" containerID="c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4" Oct 08 06:52:43 crc kubenswrapper[4810]: E1008 06:52:43.199474 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4\": container with ID starting with c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4 not found: ID does not exist" containerID="c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.199514 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4"} err="failed to get container status \"c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4\": rpc error: code = NotFound desc = could not find container \"c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4\": container with ID starting with c9df60bc0016caa57ef68d8ba15c080f7b86be5cc2f8b1191304e7ec87204cc4 not found: ID does not exist" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.199543 4810 scope.go:117] "RemoveContainer" containerID="4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84" Oct 08 06:52:43 crc kubenswrapper[4810]: E1008 06:52:43.200106 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84\": container with ID starting with 4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84 not found: ID does not exist" containerID="4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.200144 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84"} err="failed to get container status \"4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84\": rpc error: code = NotFound desc = could not find container \"4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84\": container with ID starting with 4840b44f4175537d5857fb3a69d5500b998a868daed2a7e170a3c4243ab50d84 not found: ID does not exist" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.210588 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-684c4d4db6-dr9wq"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.251911 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi29ac-account-delete-mjvf2" event={"ID":"b7b1ed43-571a-4b70-9eae-c19a7675bc59","Type":"ContainerDied","Data":"4bb120ff7e4e8942dd8b11b3aa307906a8f1066708521c268a83b6097b787a8c"} Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.251938 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi29ac-account-delete-mjvf2" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.251972 4810 scope.go:117] "RemoveContainer" containerID="d7bf8db66c158fdcb8ae6f04072efabe1937b00551aceb3c344d7a04e3d408fd" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.260801 4810 generic.go:334] "Generic (PLEG): container finished" podID="49cb68f6-1a48-4d79-90cf-34897b19dfaa" containerID="9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3" exitCode=0 Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.260869 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"49cb68f6-1a48-4d79-90cf-34897b19dfaa","Type":"ContainerDied","Data":"9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3"} Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.260899 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"49cb68f6-1a48-4d79-90cf-34897b19dfaa","Type":"ContainerDied","Data":"a49173aa83e289d210db693fd798ab7e9e14811ec7c1eecbd4f5b8356ed1f5bc"} Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.260954 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.268057 4810 generic.go:334] "Generic (PLEG): container finished" podID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerID="148c54546a54c328b2db6ab9bedc154ff074db54ed387f76af7ff0b53e27848e" exitCode=0 Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.268143 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f266b088-0be1-4d57-b8f9-c891e8ad1ed3","Type":"ContainerDied","Data":"148c54546a54c328b2db6ab9bedc154ff074db54ed387f76af7ff0b53e27848e"} Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.284003 4810 scope.go:117] "RemoveContainer" containerID="9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3" Oct 08 06:52:43 crc kubenswrapper[4810]: E1008 06:52:43.308297 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b86049c_4b80_4a1f_8e8b_51ad6e174fd7.slice/crio-conmon-70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b86049c_4b80_4a1f_8e8b_51ad6e174fd7.slice/crio-70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7486bd0_d5fd_4d4f_8596_705cc22237d5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7486bd0_d5fd_4d4f_8596_705cc22237d5.slice/crio-cb36ae6379de91ab949ca3be579c93e33e60abbc00d6023c61489cd41a69f8bf\": RecentStats: unable to find data in memory cache]" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.484327 4810 scope.go:117] "RemoveContainer" containerID="9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3" Oct 08 06:52:43 crc kubenswrapper[4810]: E1008 06:52:43.488368 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3\": container with ID starting with 9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3 not found: ID does not exist" containerID="9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.488442 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3"} err="failed to get container status \"9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3\": rpc error: code = NotFound desc = could not find container \"9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3\": container with ID starting with 9dc609b6beb42c7242fbf7d2218492e492b0cd830328066e9fd3db76ab277bc3 not found: ID does not exist" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.523915 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.553337 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.558291 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.562407 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi29ac-account-delete-mjvf2"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.566203 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi29ac-account-delete-mjvf2"] Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.640435 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-combined-ca-bundle\") pod \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.640544 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqpmt\" (UniqueName: \"kubernetes.io/projected/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-kube-api-access-fqpmt\") pod \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.640644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-config-data\") pod \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.640728 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-public-tls-certs\") pod \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.640758 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-internal-tls-certs\") pod \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.640805 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-logs\") pod \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\" (UID: \"f266b088-0be1-4d57-b8f9-c891e8ad1ed3\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.647736 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-logs" (OuterVolumeSpecName: "logs") pod "f266b088-0be1-4d57-b8f9-c891e8ad1ed3" (UID: "f266b088-0be1-4d57-b8f9-c891e8ad1ed3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.654393 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-kube-api-access-fqpmt" (OuterVolumeSpecName: "kube-api-access-fqpmt") pod "f266b088-0be1-4d57-b8f9-c891e8ad1ed3" (UID: "f266b088-0be1-4d57-b8f9-c891e8ad1ed3"). InnerVolumeSpecName "kube-api-access-fqpmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.679011 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-config-data" (OuterVolumeSpecName: "config-data") pod "f266b088-0be1-4d57-b8f9-c891e8ad1ed3" (UID: "f266b088-0be1-4d57-b8f9-c891e8ad1ed3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.715155 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f266b088-0be1-4d57-b8f9-c891e8ad1ed3" (UID: "f266b088-0be1-4d57-b8f9-c891e8ad1ed3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.718787 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f266b088-0be1-4d57-b8f9-c891e8ad1ed3" (UID: "f266b088-0be1-4d57-b8f9-c891e8ad1ed3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.732702 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.743568 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.743590 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.743598 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.743609 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqpmt\" (UniqueName: \"kubernetes.io/projected/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-kube-api-access-fqpmt\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.743617 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.756332 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f266b088-0be1-4d57-b8f9-c891e8ad1ed3" (UID: "f266b088-0be1-4d57-b8f9-c891e8ad1ed3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.844383 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-default\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.844765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8cwm\" (UniqueName: \"kubernetes.io/projected/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kube-api-access-r8cwm\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.844823 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-secrets\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.844947 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.845030 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-generated\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.845062 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-combined-ca-bundle\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.845099 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-operator-scripts\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.845128 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kolla-config\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.845149 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-galera-tls-certs\") pod \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\" (UID: \"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7\") " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.845548 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f266b088-0be1-4d57-b8f9-c891e8ad1ed3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.845824 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.846377 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.849121 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kube-api-access-r8cwm" (OuterVolumeSpecName: "kube-api-access-r8cwm") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "kube-api-access-r8cwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.849118 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.849423 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.850087 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-secrets" (OuterVolumeSpecName: "secrets") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.861912 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "mysql-db") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.879510 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.899859 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" (UID: "0b86049c-4b80-4a1f-8e8b-51ad6e174fd7"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.946896 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.946936 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.946948 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.946956 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.946982 4810 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.946991 4810 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.946999 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.947007 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8cwm\" (UniqueName: \"kubernetes.io/projected/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-kube-api-access-r8cwm\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.947015 4810 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:43 crc kubenswrapper[4810]: I1008 06:52:43.961632 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.019737 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053028 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-tls\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-plugins\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053300 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053350 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-pod-info\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053459 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053525 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-server-conf\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053560 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-confd\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053596 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-erlang-cookie-secret\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053622 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-plugins-conf\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053686 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-erlang-cookie\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.053760 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm62m\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-kube-api-access-xm62m\") pod \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\" (UID: \"2a44c96c-b96c-409c-9c9e-c049d9fe68b5\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.054341 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.055646 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.055764 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.059266 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.061495 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.061498 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.062072 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.062139 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-pod-info" (OuterVolumeSpecName: "pod-info") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.063085 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-kube-api-access-xm62m" (OuterVolumeSpecName: "kube-api-access-xm62m") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "kube-api-access-xm62m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.093001 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" path="/var/lib/kubelet/pods/016c35ce-3072-485f-86b8-bbe927f78f81/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.095194 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" path="/var/lib/kubelet/pods/06237db0-424d-4bd7-ae4f-3f7af9cc4f92/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.096058 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" path="/var/lib/kubelet/pods/12678e95-0510-4cd6-86c8-701ea97a247e/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.097518 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" path="/var/lib/kubelet/pods/30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.098162 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e33182-d302-417e-8e4e-fec4808047a4" path="/var/lib/kubelet/pods/35e33182-d302-417e-8e4e-fec4808047a4/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.099023 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" path="/var/lib/kubelet/pods/48ca6de6-470e-4226-b67b-e58bf64415e3/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.100388 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49cb68f6-1a48-4d79-90cf-34897b19dfaa" path="/var/lib/kubelet/pods/49cb68f6-1a48-4d79-90cf-34897b19dfaa/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.101008 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64350cd1-38b8-4c64-8c01-105432d1b91f" path="/var/lib/kubelet/pods/64350cd1-38b8-4c64-8c01-105432d1b91f/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.101724 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" path="/var/lib/kubelet/pods/8832ab0b-b264-4db9-8b70-87920c1e4826/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.103181 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c1351c1-21aa-4004-a402-a603343c58d7" path="/var/lib/kubelet/pods/9c1351c1-21aa-4004-a402-a603343c58d7/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.103896 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a940c7a1-b3cc-4168-99e8-750c87845314" path="/var/lib/kubelet/pods/a940c7a1-b3cc-4168-99e8-750c87845314/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.104562 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b1ed43-571a-4b70-9eae-c19a7675bc59" path="/var/lib/kubelet/pods/b7b1ed43-571a-4b70-9eae-c19a7675bc59/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.105727 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" path="/var/lib/kubelet/pods/d7486bd0-d5fd-4d4f-8596-705cc22237d5/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.106498 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" path="/var/lib/kubelet/pods/dced1083-d856-4f53-bb6e-19fc64ba95a5/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.107982 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebbf3e4c-75db-4d41-8220-4b5c687d815f" path="/var/lib/kubelet/pods/ebbf3e4c-75db-4d41-8220-4b5c687d815f/volumes" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.114378 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-server-conf" (OuterVolumeSpecName: "server-conf") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.119816 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data" (OuterVolumeSpecName: "config-data") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.154136 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2a44c96c-b96c-409c-9c9e-c049d9fe68b5" (UID: "2a44c96c-b96c-409c-9c9e-c049d9fe68b5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159008 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm62m\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-kube-api-access-xm62m\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159317 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159367 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159413 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159428 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159446 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159460 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159472 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159487 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159500 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.159513 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2a44c96c-b96c-409c-9c9e-c049d9fe68b5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.176078 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.180114 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7/ovn-northd/0.log" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.180211 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.206800 4810 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 08 06:52:44 crc kubenswrapper[4810]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-08T06:52:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 08 06:52:44 crc kubenswrapper[4810]: /etc/init.d/functions: line 589: 393 Alarm clock "$@" Oct 08 06:52:44 crc kubenswrapper[4810]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-qtj9x" message=< Oct 08 06:52:44 crc kubenswrapper[4810]: Exiting ovn-controller (1) [FAILED] Oct 08 06:52:44 crc kubenswrapper[4810]: Killing ovn-controller (1) [ OK ] Oct 08 06:52:44 crc kubenswrapper[4810]: Killing ovn-controller (1) with SIGKILL [ OK ] Oct 08 06:52:44 crc kubenswrapper[4810]: 2025-10-08T06:52:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 08 06:52:44 crc kubenswrapper[4810]: /etc/init.d/functions: line 589: 393 Alarm clock "$@" Oct 08 06:52:44 crc kubenswrapper[4810]: > Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.206837 4810 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 08 06:52:44 crc kubenswrapper[4810]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-08T06:52:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 08 06:52:44 crc kubenswrapper[4810]: /etc/init.d/functions: line 589: 393 Alarm clock "$@" Oct 08 06:52:44 crc kubenswrapper[4810]: > pod="openstack/ovn-controller-qtj9x" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" containerID="cri-o://3110ad491e155ae79f17fbaedfb3c85a0be30ad15f2ba193fc7eb67f4b9b78a1" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.206895 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-qtj9x" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" containerID="cri-o://3110ad491e155ae79f17fbaedfb3c85a0be30ad15f2ba193fc7eb67f4b9b78a1" gracePeriod=22 Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.222547 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.235235 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.237017 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.237065 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" containerName="nova-cell1-conductor-conductor" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.262437 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-scripts\") pod \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.262529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-rundir\") pod \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.262701 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-northd-tls-certs\") pod \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.262843 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-combined-ca-bundle\") pod \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.262925 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-config\") pod \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.263089 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp7gp\" (UniqueName: \"kubernetes.io/projected/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-kube-api-access-mp7gp\") pod \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.263118 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-metrics-certs-tls-certs\") pod \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\" (UID: \"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.264177 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.264626 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-config" (OuterVolumeSpecName: "config") pod "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" (UID: "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.266506 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" (UID: "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.268469 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-kube-api-access-mp7gp" (OuterVolumeSpecName: "kube-api-access-mp7gp") pod "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" (UID: "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7"). InnerVolumeSpecName "kube-api-access-mp7gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.279321 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-scripts" (OuterVolumeSpecName: "scripts") pod "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" (UID: "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.287300 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" (UID: "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.314091 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f266b088-0be1-4d57-b8f9-c891e8ad1ed3","Type":"ContainerDied","Data":"ecf756d03b53ec62326985bad7be058e4218947efde29ea2cde6f0ae21d613e5"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.315042 4810 scope.go:117] "RemoveContainer" containerID="148c54546a54c328b2db6ab9bedc154ff074db54ed387f76af7ff0b53e27848e" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.315483 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.327344 4810 generic.go:334] "Generic (PLEG): container finished" podID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerID="5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98" exitCode=0 Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.327446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2a44c96c-b96c-409c-9c9e-c049d9fe68b5","Type":"ContainerDied","Data":"5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.327479 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2a44c96c-b96c-409c-9c9e-c049d9fe68b5","Type":"ContainerDied","Data":"708b0993f45847c122345d0e8e83973a68b9edfdee190a0cbc4bc091346d9761"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.327568 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.340555 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-qtj9x_111360ca-3afe-4eb6-83a0-c60ca579c82b/ovn-controller/0.log" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.340637 4810 generic.go:334] "Generic (PLEG): container finished" podID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerID="3110ad491e155ae79f17fbaedfb3c85a0be30ad15f2ba193fc7eb67f4b9b78a1" exitCode=137 Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.340767 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x" event={"ID":"111360ca-3afe-4eb6-83a0-c60ca579c82b","Type":"ContainerDied","Data":"3110ad491e155ae79f17fbaedfb3c85a0be30ad15f2ba193fc7eb67f4b9b78a1"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.345790 4810 generic.go:334] "Generic (PLEG): container finished" podID="4deb0612-547f-4067-b95b-5794663d21aa" containerID="2d507eb2a2ecc2839157b138e7210f5bbddda79641955e41f82171d217de786c" exitCode=0 Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.345940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4deb0612-547f-4067-b95b-5794663d21aa","Type":"ContainerDied","Data":"2d507eb2a2ecc2839157b138e7210f5bbddda79641955e41f82171d217de786c"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.348685 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7/ovn-northd/0.log" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.348748 4810 generic.go:334] "Generic (PLEG): container finished" podID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" exitCode=139 Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.348819 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.348858 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7","Type":"ContainerDied","Data":"a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.348892 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7","Type":"ContainerDied","Data":"ee3158257410582bd8eae3a8191820076a2b25b91711a64ef3ac1ac03ae953d1"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.351205 4810 generic.go:334] "Generic (PLEG): container finished" podID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerID="70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb" exitCode=0 Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.351262 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7","Type":"ContainerDied","Data":"70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.351290 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0b86049c-4b80-4a1f-8e8b-51ad6e174fd7","Type":"ContainerDied","Data":"706c6128ecbaea4077f4cb0e88470ac3c5e4ee2303ca3ba4fa21eb87439e4fc4"} Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.351347 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.356393 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" (UID: "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.366055 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.366729 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.366741 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp7gp\" (UniqueName: \"kubernetes.io/projected/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-kube-api-access-mp7gp\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.366750 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.366761 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.366770 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.366055 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" (UID: "7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.453939 4810 scope.go:117] "RemoveContainer" containerID="744e6bf0d1c62016c3949d6f7346fe561790ee47cbd2b70695cea07476938ff0" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.461500 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.468581 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.484231 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.492559 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.501006 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.507571 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.507635 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.509554 4810 scope.go:117] "RemoveContainer" containerID="5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.559130 4810 scope.go:117] "RemoveContainer" containerID="a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.620705 4810 scope.go:117] "RemoveContainer" containerID="5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.624101 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98\": container with ID starting with 5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98 not found: ID does not exist" containerID="5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.624145 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98"} err="failed to get container status \"5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98\": rpc error: code = NotFound desc = could not find container \"5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98\": container with ID starting with 5ea9e9284a52e9ef7b97a43cc872bf116948d350eab6027fa9f087aa2274fc98 not found: ID does not exist" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.624172 4810 scope.go:117] "RemoveContainer" containerID="a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.625098 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1\": container with ID starting with a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1 not found: ID does not exist" containerID="a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.625141 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1"} err="failed to get container status \"a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1\": rpc error: code = NotFound desc = could not find container \"a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1\": container with ID starting with a725f102d9f75b13832d35042780667df4fe72933e63af67d31e666d345dfdb1 not found: ID does not exist" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.625176 4810 scope.go:117] "RemoveContainer" containerID="6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.703537 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-qtj9x_111360ca-3afe-4eb6-83a0-c60ca579c82b/ovn-controller/0.log" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.703620 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.723530 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.724721 4810 scope.go:117] "RemoveContainer" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.726467 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.751146 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.755460 4810 scope.go:117] "RemoveContainer" containerID="6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.756931 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b\": container with ID starting with 6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b not found: ID does not exist" containerID="6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.757004 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b"} err="failed to get container status \"6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b\": rpc error: code = NotFound desc = could not find container \"6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b\": container with ID starting with 6133cb1d06b2d4ae03c4923f664ca349191fe52f2b49f560e72f841989fce86b not found: ID does not exist" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.757042 4810 scope.go:117] "RemoveContainer" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.757274 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c\": container with ID starting with a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c not found: ID does not exist" containerID="a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.757305 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c"} err="failed to get container status \"a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c\": rpc error: code = NotFound desc = could not find container \"a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c\": container with ID starting with a1ec9485a27e150314ee0d9a0344d2a4930549782a25de63e3e030971ae7574c not found: ID does not exist" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.757324 4810 scope.go:117] "RemoveContainer" containerID="70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.758662 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.778374 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-combined-ca-bundle\") pod \"111360ca-3afe-4eb6-83a0-c60ca579c82b\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.778523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run\") pod \"111360ca-3afe-4eb6-83a0-c60ca579c82b\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.778549 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-log-ovn\") pod \"111360ca-3afe-4eb6-83a0-c60ca579c82b\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.778590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111360ca-3afe-4eb6-83a0-c60ca579c82b-scripts\") pod \"111360ca-3afe-4eb6-83a0-c60ca579c82b\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.778670 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run-ovn\") pod \"111360ca-3afe-4eb6-83a0-c60ca579c82b\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.778705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55m4\" (UniqueName: \"kubernetes.io/projected/111360ca-3afe-4eb6-83a0-c60ca579c82b-kube-api-access-x55m4\") pod \"111360ca-3afe-4eb6-83a0-c60ca579c82b\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.778768 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-ovn-controller-tls-certs\") pod \"111360ca-3afe-4eb6-83a0-c60ca579c82b\" (UID: \"111360ca-3afe-4eb6-83a0-c60ca579c82b\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.787153 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run" (OuterVolumeSpecName: "var-run") pod "111360ca-3afe-4eb6-83a0-c60ca579c82b" (UID: "111360ca-3afe-4eb6-83a0-c60ca579c82b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.787173 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "111360ca-3afe-4eb6-83a0-c60ca579c82b" (UID: "111360ca-3afe-4eb6-83a0-c60ca579c82b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.787255 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "111360ca-3afe-4eb6-83a0-c60ca579c82b" (UID: "111360ca-3afe-4eb6-83a0-c60ca579c82b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.790905 4810 scope.go:117] "RemoveContainer" containerID="1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.791144 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111360ca-3afe-4eb6-83a0-c60ca579c82b-kube-api-access-x55m4" (OuterVolumeSpecName: "kube-api-access-x55m4") pod "111360ca-3afe-4eb6-83a0-c60ca579c82b" (UID: "111360ca-3afe-4eb6-83a0-c60ca579c82b"). InnerVolumeSpecName "kube-api-access-x55m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.802347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111360ca-3afe-4eb6-83a0-c60ca579c82b-scripts" (OuterVolumeSpecName: "scripts") pod "111360ca-3afe-4eb6-83a0-c60ca579c82b" (UID: "111360ca-3afe-4eb6-83a0-c60ca579c82b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.838720 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "111360ca-3afe-4eb6-83a0-c60ca579c82b" (UID: "111360ca-3afe-4eb6-83a0-c60ca579c82b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.852924 4810 scope.go:117] "RemoveContainer" containerID="70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.854446 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb\": container with ID starting with 70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb not found: ID does not exist" containerID="70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.854506 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb"} err="failed to get container status \"70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb\": rpc error: code = NotFound desc = could not find container \"70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb\": container with ID starting with 70e828209f0c0d2143401712b7e6efeb6970101cc819b9a9d3ddec5dbabf23cb not found: ID does not exist" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.854553 4810 scope.go:117] "RemoveContainer" containerID="1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.855208 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b\": container with ID starting with 1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b not found: ID does not exist" containerID="1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.855280 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b"} err="failed to get container status \"1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b\": rpc error: code = NotFound desc = could not find container \"1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b\": container with ID starting with 1f0fcf5e9ab50f9f2464589ce77cabba43a0773960da07b791216b926b72ec3b not found: ID does not exist" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880401 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880482 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt45m\" (UniqueName: \"kubernetes.io/projected/493810ad-f07f-4871-8063-5d54cd5786ea-kube-api-access-rt45m\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880525 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880550 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-server-conf\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880595 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4deb0612-547f-4067-b95b-5794663d21aa-erlang-cookie-secret\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-plugins-conf\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880664 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-scripts\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880680 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-public-tls-certs\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880711 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-plugins\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880738 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4deb0612-547f-4067-b95b-5794663d21aa-pod-info\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880754 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhz55\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-kube-api-access-zhz55\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880772 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-fernet-keys\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-credential-keys\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880812 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-config-data\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880835 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-erlang-cookie\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880860 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-tls\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880884 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880903 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-confd\") pod \"4deb0612-547f-4067-b95b-5794663d21aa\" (UID: \"4deb0612-547f-4067-b95b-5794663d21aa\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.880923 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-combined-ca-bundle\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.881239 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.881255 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.881264 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.881272 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111360ca-3afe-4eb6-83a0-c60ca579c82b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.881281 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/111360ca-3afe-4eb6-83a0-c60ca579c82b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.881289 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55m4\" (UniqueName: \"kubernetes.io/projected/111360ca-3afe-4eb6-83a0-c60ca579c82b-kube-api-access-x55m4\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.883555 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.883849 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.884364 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.884483 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-kube-api-access-zhz55" (OuterVolumeSpecName: "kube-api-access-zhz55") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "kube-api-access-zhz55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.886021 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.887624 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493810ad-f07f-4871-8063-5d54cd5786ea-kube-api-access-rt45m" (OuterVolumeSpecName: "kube-api-access-rt45m") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "kube-api-access-rt45m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.889598 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.890184 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4deb0612-547f-4067-b95b-5794663d21aa-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.890194 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-scripts" (OuterVolumeSpecName: "scripts") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.890744 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4deb0612-547f-4067-b95b-5794663d21aa-pod-info" (OuterVolumeSpecName: "pod-info") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.890809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.898743 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.903102 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "111360ca-3afe-4eb6-83a0-c60ca579c82b" (UID: "111360ca-3afe-4eb6-83a0-c60ca579c82b"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.913256 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data" (OuterVolumeSpecName: "config-data") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.926163 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.929196 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-config-data" (OuterVolumeSpecName: "config-data") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: E1008 06:52:44.936154 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs podName:493810ad-f07f-4871-8063-5d54cd5786ea nodeName:}" failed. No retries permitted until 2025-10-08 06:52:45.436117561 +0000 UTC m=+1268.070557381 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "internal-tls-certs" (UniqueName: "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea") : error deleting /var/lib/kubelet/pods/493810ad-f07f-4871-8063-5d54cd5786ea/volume-subpaths: remove /var/lib/kubelet/pods/493810ad-f07f-4871-8063-5d54cd5786ea/volume-subpaths: no such file or directory Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.938391 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.952500 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-server-conf" (OuterVolumeSpecName: "server-conf") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983662 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983730 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt45m\" (UniqueName: \"kubernetes.io/projected/493810ad-f07f-4871-8063-5d54cd5786ea-kube-api-access-rt45m\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983748 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983761 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983774 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/111360ca-3afe-4eb6-83a0-c60ca579c82b-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983785 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4deb0612-547f-4067-b95b-5794663d21aa-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983795 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4deb0612-547f-4067-b95b-5794663d21aa-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983806 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983817 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983830 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983839 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4deb0612-547f-4067-b95b-5794663d21aa-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983850 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhz55\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-kube-api-access-zhz55\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983862 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983872 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983882 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983892 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983902 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:44 crc kubenswrapper[4810]: I1008 06:52:44.983912 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.005836 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4deb0612-547f-4067-b95b-5794663d21aa" (UID: "4deb0612-547f-4067-b95b-5794663d21aa"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.014177 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.085360 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.085408 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4deb0612-547f-4067-b95b-5794663d21aa-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.297068 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.380660 4810 generic.go:334] "Generic (PLEG): container finished" podID="493810ad-f07f-4871-8063-5d54cd5786ea" containerID="5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e" exitCode=0 Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.380741 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b4dd4748f-7h7bc" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.380775 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b4dd4748f-7h7bc" event={"ID":"493810ad-f07f-4871-8063-5d54cd5786ea","Type":"ContainerDied","Data":"5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e"} Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.380827 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b4dd4748f-7h7bc" event={"ID":"493810ad-f07f-4871-8063-5d54cd5786ea","Type":"ContainerDied","Data":"c18483cfeeab72004f52e63b3aa0a878e963ab4fa03645b5570b9c2dd3b3aa1f"} Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.380850 4810 scope.go:117] "RemoveContainer" containerID="5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.388887 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-combined-ca-bundle\") pod \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.388938 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data-custom\") pod \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.389016 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data\") pod \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.389139 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-logs\") pod \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.389175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9djf\" (UniqueName: \"kubernetes.io/projected/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-kube-api-access-k9djf\") pod \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\" (UID: \"aabc3b24-1971-4326-aff0-12ddbf1f5ea1\") " Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.390744 4810 generic.go:334] "Generic (PLEG): container finished" podID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerID="833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b" exitCode=0 Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.390851 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" event={"ID":"aabc3b24-1971-4326-aff0-12ddbf1f5ea1","Type":"ContainerDied","Data":"833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b"} Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.390888 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" event={"ID":"aabc3b24-1971-4326-aff0-12ddbf1f5ea1","Type":"ContainerDied","Data":"71873762cf344d06269eb1a0804ef2572b9f9c16eda30852737b8d83a5a82a1e"} Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.390992 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4bcc558b-zx56q" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.391488 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-logs" (OuterVolumeSpecName: "logs") pod "aabc3b24-1971-4326-aff0-12ddbf1f5ea1" (UID: "aabc3b24-1971-4326-aff0-12ddbf1f5ea1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.402539 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aabc3b24-1971-4326-aff0-12ddbf1f5ea1" (UID: "aabc3b24-1971-4326-aff0-12ddbf1f5ea1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.402664 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-kube-api-access-k9djf" (OuterVolumeSpecName: "kube-api-access-k9djf") pod "aabc3b24-1971-4326-aff0-12ddbf1f5ea1" (UID: "aabc3b24-1971-4326-aff0-12ddbf1f5ea1"). InnerVolumeSpecName "kube-api-access-k9djf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.414822 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-qtj9x_111360ca-3afe-4eb6-83a0-c60ca579c82b/ovn-controller/0.log" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.415009 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qtj9x" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.415490 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qtj9x" event={"ID":"111360ca-3afe-4eb6-83a0-c60ca579c82b","Type":"ContainerDied","Data":"43d74d5e2fe2ea4b1d6bd07651d35dd091c2d05f4173f63194d7ff518c2a7bf3"} Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.417836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4deb0612-547f-4067-b95b-5794663d21aa","Type":"ContainerDied","Data":"0bd79405f9b24e71b2a92fb1817c23cc334b58a5d026f67058ed9df15150c29d"} Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.417912 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.427126 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aabc3b24-1971-4326-aff0-12ddbf1f5ea1" (UID: "aabc3b24-1971-4326-aff0-12ddbf1f5ea1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.450991 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data" (OuterVolumeSpecName: "config-data") pod "aabc3b24-1971-4326-aff0-12ddbf1f5ea1" (UID: "aabc3b24-1971-4326-aff0-12ddbf1f5ea1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.490440 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs\") pod \"493810ad-f07f-4871-8063-5d54cd5786ea\" (UID: \"493810ad-f07f-4871-8063-5d54cd5786ea\") " Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.490765 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.490784 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-logs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.490795 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9djf\" (UniqueName: \"kubernetes.io/projected/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-kube-api-access-k9djf\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.490807 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.490815 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aabc3b24-1971-4326-aff0-12ddbf1f5ea1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.494825 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "493810ad-f07f-4871-8063-5d54cd5786ea" (UID: "493810ad-f07f-4871-8063-5d54cd5786ea"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.556377 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.573346 4810 scope.go:117] "RemoveContainer" containerID="5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e" Oct 08 06:52:45 crc kubenswrapper[4810]: E1008 06:52:45.574290 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e\": container with ID starting with 5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e not found: ID does not exist" containerID="5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.574446 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e"} err="failed to get container status \"5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e\": rpc error: code = NotFound desc = could not find container \"5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e\": container with ID starting with 5068fea3ae181ecc20ea421c44813cf7775b769f8da78a4da6da78ce3d71d82e not found: ID does not exist" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.574571 4810 scope.go:117] "RemoveContainer" containerID="833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.583012 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.595649 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-qtj9x"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.596920 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493810ad-f07f-4871-8063-5d54cd5786ea-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.608743 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-qtj9x"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.608872 4810 scope.go:117] "RemoveContainer" containerID="585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.628756 4810 scope.go:117] "RemoveContainer" containerID="833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b" Oct 08 06:52:45 crc kubenswrapper[4810]: E1008 06:52:45.630696 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b\": container with ID starting with 833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b not found: ID does not exist" containerID="833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.630732 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b"} err="failed to get container status \"833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b\": rpc error: code = NotFound desc = could not find container \"833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b\": container with ID starting with 833b992035b3474484f0df68909f0d01397eb1b38768522de6db044ee6ae3f7b not found: ID does not exist" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.630762 4810 scope.go:117] "RemoveContainer" containerID="585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f" Oct 08 06:52:45 crc kubenswrapper[4810]: E1008 06:52:45.631213 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f\": container with ID starting with 585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f not found: ID does not exist" containerID="585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.631242 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f"} err="failed to get container status \"585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f\": rpc error: code = NotFound desc = could not find container \"585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f\": container with ID starting with 585ab98d97e28532f44e901091a0e0dadfc7f2ef36a7ab406b47a9f30b8a3f9f not found: ID does not exist" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.631261 4810 scope.go:117] "RemoveContainer" containerID="3110ad491e155ae79f17fbaedfb3c85a0be30ad15f2ba193fc7eb67f4b9b78a1" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.711408 4810 scope.go:117] "RemoveContainer" containerID="2d507eb2a2ecc2839157b138e7210f5bbddda79641955e41f82171d217de786c" Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.722898 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6b4dd4748f-7h7bc"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.726580 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6b4dd4748f-7h7bc"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.740349 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5b4bcc558b-zx56q"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.745434 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-5b4bcc558b-zx56q"] Oct 08 06:52:45 crc kubenswrapper[4810]: I1008 06:52:45.756412 4810 scope.go:117] "RemoveContainer" containerID="c3755a82f85d0942937435d7ed4103b19749825c63d0c7188863bb653199761b" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.010884 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.089482 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" path="/var/lib/kubelet/pods/0b86049c-4b80-4a1f-8e8b-51ad6e174fd7/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.090298 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" path="/var/lib/kubelet/pods/111360ca-3afe-4eb6-83a0-c60ca579c82b/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.091545 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" path="/var/lib/kubelet/pods/2a44c96c-b96c-409c-9c9e-c049d9fe68b5/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.093380 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493810ad-f07f-4871-8063-5d54cd5786ea" path="/var/lib/kubelet/pods/493810ad-f07f-4871-8063-5d54cd5786ea/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.094307 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4deb0612-547f-4067-b95b-5794663d21aa" path="/var/lib/kubelet/pods/4deb0612-547f-4067-b95b-5794663d21aa/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.095654 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" path="/var/lib/kubelet/pods/7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.096255 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" path="/var/lib/kubelet/pods/aabc3b24-1971-4326-aff0-12ddbf1f5ea1/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.096882 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" path="/var/lib/kubelet/pods/f266b088-0be1-4d57-b8f9-c891e8ad1ed3/volumes" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.104941 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xp5b\" (UniqueName: \"kubernetes.io/projected/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-kube-api-access-2xp5b\") pod \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.105013 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-config-data\") pod \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.105119 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-combined-ca-bundle\") pod \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\" (UID: \"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.109181 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-kube-api-access-2xp5b" (OuterVolumeSpecName: "kube-api-access-2xp5b") pod "e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" (UID: "e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b"). InnerVolumeSpecName "kube-api-access-2xp5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.110884 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.116873 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.132660 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-config-data" (OuterVolumeSpecName: "config-data") pod "e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" (UID: "e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.135991 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" (UID: "e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.206716 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-config-data\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.206763 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-scripts\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.206798 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-sg-core-conf-yaml\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.206829 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-ceilometer-tls-certs\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.206886 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-run-httpd\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.206939 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wqcf\" (UniqueName: \"kubernetes.io/projected/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-kube-api-access-8wqcf\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.206995 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-combined-ca-bundle\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.207020 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-log-httpd\") pod \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\" (UID: \"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d\") " Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.207313 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.207325 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xp5b\" (UniqueName: \"kubernetes.io/projected/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-kube-api-access-2xp5b\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.207335 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.207524 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.208811 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.211160 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-kube-api-access-8wqcf" (OuterVolumeSpecName: "kube-api-access-8wqcf") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "kube-api-access-8wqcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.216264 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-scripts" (OuterVolumeSpecName: "scripts") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.227450 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.249118 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.264344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.306792 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-config-data" (OuterVolumeSpecName: "config-data") pod "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" (UID: "b6d58cc0-e60f-49c7-ad4c-d4db94cea14d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308647 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308679 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308689 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308700 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308709 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308719 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wqcf\" (UniqueName: \"kubernetes.io/projected/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-kube-api-access-8wqcf\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308727 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.308734 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.427355 4810 generic.go:334] "Generic (PLEG): container finished" podID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" exitCode=0 Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.427426 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.427430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b","Type":"ContainerDied","Data":"f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e"} Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.427535 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b","Type":"ContainerDied","Data":"b8dd244f52cb0ecb3525951046b786fa3edfb96ad12d542da4ee883c31710832"} Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.427554 4810 scope.go:117] "RemoveContainer" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.435763 4810 generic.go:334] "Generic (PLEG): container finished" podID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerID="6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c" exitCode=0 Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.435826 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerDied","Data":"6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c"} Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.435852 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b6d58cc0-e60f-49c7-ad4c-d4db94cea14d","Type":"ContainerDied","Data":"96aa368956d59be4b085c7e5b85434c43547f5720ca062f547c74939c7f33876"} Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.436067 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.466893 4810 scope.go:117] "RemoveContainer" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" Oct 08 06:52:46 crc kubenswrapper[4810]: E1008 06:52:46.467518 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e\": container with ID starting with f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e not found: ID does not exist" containerID="f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.467559 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e"} err="failed to get container status \"f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e\": rpc error: code = NotFound desc = could not find container \"f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e\": container with ID starting with f3e481db024743dbccd5827f59f4509d1fdcd1669158876de034dd61fb1d707e not found: ID does not exist" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.467588 4810 scope.go:117] "RemoveContainer" containerID="9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.469361 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.474508 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.493783 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.500160 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.506367 4810 scope.go:117] "RemoveContainer" containerID="9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.531771 4810 scope.go:117] "RemoveContainer" containerID="6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.566953 4810 scope.go:117] "RemoveContainer" containerID="1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.600200 4810 scope.go:117] "RemoveContainer" containerID="9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc" Oct 08 06:52:46 crc kubenswrapper[4810]: E1008 06:52:46.601032 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc\": container with ID starting with 9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc not found: ID does not exist" containerID="9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.601092 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc"} err="failed to get container status \"9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc\": rpc error: code = NotFound desc = could not find container \"9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc\": container with ID starting with 9c85bb69e376dc3508c636fe054f930afacdf1a5e93b28181fca72de0e9515cc not found: ID does not exist" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.601128 4810 scope.go:117] "RemoveContainer" containerID="9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049" Oct 08 06:52:46 crc kubenswrapper[4810]: E1008 06:52:46.601900 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049\": container with ID starting with 9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049 not found: ID does not exist" containerID="9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.601938 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049"} err="failed to get container status \"9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049\": rpc error: code = NotFound desc = could not find container \"9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049\": container with ID starting with 9f66da48f6fd28c33d61088c1c1c8566e4321b9c582e9a5394b756517c193049 not found: ID does not exist" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.601971 4810 scope.go:117] "RemoveContainer" containerID="6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c" Oct 08 06:52:46 crc kubenswrapper[4810]: E1008 06:52:46.602618 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c\": container with ID starting with 6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c not found: ID does not exist" containerID="6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.602688 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c"} err="failed to get container status \"6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c\": rpc error: code = NotFound desc = could not find container \"6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c\": container with ID starting with 6ae55194fcacb26ab9b9a5724d001d7dd71a284dfc6f8676a63c55900db48f7c not found: ID does not exist" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.602730 4810 scope.go:117] "RemoveContainer" containerID="1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495" Oct 08 06:52:46 crc kubenswrapper[4810]: E1008 06:52:46.603377 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495\": container with ID starting with 1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495 not found: ID does not exist" containerID="1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.603434 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495"} err="failed to get container status \"1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495\": rpc error: code = NotFound desc = could not find container \"1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495\": container with ID starting with 1629095e78f131dca6ab18d301ebd20e05449451f06523fdb05a79adae991495 not found: ID does not exist" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.898223 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 06:52:46 crc kubenswrapper[4810]: I1008 06:52:46.898270 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.267643 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.268192 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.268578 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.268648 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.269462 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.270608 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.272434 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:47 crc kubenswrapper[4810]: E1008 06:52:47.272483 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:52:48 crc kubenswrapper[4810]: I1008 06:52:48.084451 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" path="/var/lib/kubelet/pods/b6d58cc0-e60f-49c7-ad4c-d4db94cea14d/volumes" Oct 08 06:52:48 crc kubenswrapper[4810]: I1008 06:52:48.085416 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" path="/var/lib/kubelet/pods/e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b/volumes" Oct 08 06:52:49 crc kubenswrapper[4810]: I1008 06:52:49.431883 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:52:49 crc kubenswrapper[4810]: I1008 06:52:49.432013 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.268202 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.268887 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.269496 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.269821 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.269864 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.272315 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.274519 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:52 crc kubenswrapper[4810]: E1008 06:52:52.283913 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.267637 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.270601 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.270625 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.271257 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.271369 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.272391 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.274138 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:52:57 crc kubenswrapper[4810]: E1008 06:52:57.274265 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.152500 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.221703 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk47z\" (UniqueName: \"kubernetes.io/projected/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-kube-api-access-hk47z\") pod \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.221752 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-combined-ca-bundle\") pod \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.221827 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-httpd-config\") pod \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.221887 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-config\") pod \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.221922 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-public-tls-certs\") pod \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.221942 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-internal-tls-certs\") pod \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.222016 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-ovndb-tls-certs\") pod \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\" (UID: \"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0\") " Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.227065 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-kube-api-access-hk47z" (OuterVolumeSpecName: "kube-api-access-hk47z") pod "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" (UID: "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0"). InnerVolumeSpecName "kube-api-access-hk47z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.228549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" (UID: "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.257903 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" (UID: "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.258164 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" (UID: "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.258551 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-config" (OuterVolumeSpecName: "config") pod "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" (UID: "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.259157 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" (UID: "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.274496 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" (UID: "6d95dc4d-f167-494a-b09f-09b5e0cf8ab0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.323854 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.323893 4810 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.323903 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk47z\" (UniqueName: \"kubernetes.io/projected/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-kube-api-access-hk47z\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.323913 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.323924 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.323932 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-config\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.323940 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.602917 4810 generic.go:334] "Generic (PLEG): container finished" podID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerID="77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd" exitCode=0 Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.602957 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f65ff9-7hhz2" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.602982 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f65ff9-7hhz2" event={"ID":"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0","Type":"ContainerDied","Data":"77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd"} Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.603031 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f65ff9-7hhz2" event={"ID":"6d95dc4d-f167-494a-b09f-09b5e0cf8ab0","Type":"ContainerDied","Data":"d131e54d6c7d758c981b65b803d5b97e5dd547f70afde36f4dcf40a8ff1cae31"} Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.603053 4810 scope.go:117] "RemoveContainer" containerID="a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.643914 4810 scope.go:117] "RemoveContainer" containerID="77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.649216 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6f65ff9-7hhz2"] Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.658011 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6f65ff9-7hhz2"] Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.668879 4810 scope.go:117] "RemoveContainer" containerID="a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df" Oct 08 06:52:59 crc kubenswrapper[4810]: E1008 06:52:59.669479 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df\": container with ID starting with a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df not found: ID does not exist" containerID="a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.669545 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df"} err="failed to get container status \"a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df\": rpc error: code = NotFound desc = could not find container \"a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df\": container with ID starting with a3b44ddfdb542acb5fbd5099daf6543ba2e53a3f5bd8db6cacee43bb901ec4df not found: ID does not exist" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.669579 4810 scope.go:117] "RemoveContainer" containerID="77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd" Oct 08 06:52:59 crc kubenswrapper[4810]: E1008 06:52:59.671017 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd\": container with ID starting with 77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd not found: ID does not exist" containerID="77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd" Oct 08 06:52:59 crc kubenswrapper[4810]: I1008 06:52:59.671055 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd"} err="failed to get container status \"77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd\": rpc error: code = NotFound desc = could not find container \"77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd\": container with ID starting with 77ffe8484b5e89dc734c1b1f87af92c1af72812d3d2b2dd40d21d004938300cd not found: ID does not exist" Oct 08 06:53:00 crc kubenswrapper[4810]: I1008 06:53:00.088669 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" path="/var/lib/kubelet/pods/6d95dc4d-f167-494a-b09f-09b5e0cf8ab0/volumes" Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.267643 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.268525 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.269124 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.269186 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.269456 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.271287 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.272997 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 06:53:02 crc kubenswrapper[4810]: E1008 06:53:02.273051 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xlsjm" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.189325 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xlsjm_fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37/ovs-vswitchd/0.log" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.190335 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.331764 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hftz\" (UniqueName: \"kubernetes.io/projected/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-kube-api-access-6hftz\") pod \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.331848 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-run\") pod \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.331930 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-etc-ovs\") pod \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.331988 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-log\") pod \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332034 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-scripts\") pod \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332073 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-lib\") pod \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\" (UID: \"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37\") " Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332151 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" (UID: "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-lib" (OuterVolumeSpecName: "var-lib") pod "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" (UID: "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332302 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-run" (OuterVolumeSpecName: "var-run") pod "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" (UID: "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332340 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-log" (OuterVolumeSpecName: "var-log") pod "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" (UID: "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332609 4810 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332626 4810 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-log\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332638 4810 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-lib\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.332649 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.333413 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-scripts" (OuterVolumeSpecName: "scripts") pod "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" (UID: "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.339436 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-kube-api-access-6hftz" (OuterVolumeSpecName: "kube-api-access-6hftz") pod "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" (UID: "fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37"). InnerVolumeSpecName "kube-api-access-6hftz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.434401 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hftz\" (UniqueName: \"kubernetes.io/projected/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-kube-api-access-6hftz\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.434541 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.680810 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xlsjm_fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37/ovs-vswitchd/0.log" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.682075 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" exitCode=137 Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.682079 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xlsjm" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.682107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xlsjm" event={"ID":"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37","Type":"ContainerDied","Data":"785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611"} Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.682572 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xlsjm" event={"ID":"fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37","Type":"ContainerDied","Data":"95eab66a50345cf0bb53a0a140c2990323fab8df4865001953b83fd3efd14cd9"} Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.682619 4810 scope.go:117] "RemoveContainer" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.700606 4810 generic.go:334] "Generic (PLEG): container finished" podID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerID="758c08848dcf56803d3340df9307815df8fed52c268f9d576bd87dce38f15071" exitCode=137 Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.700656 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"758c08848dcf56803d3340df9307815df8fed52c268f9d576bd87dce38f15071"} Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.722111 4810 scope.go:117] "RemoveContainer" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.733161 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-xlsjm"] Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.741044 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-xlsjm"] Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.747948 4810 scope.go:117] "RemoveContainer" containerID="5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.784321 4810 scope.go:117] "RemoveContainer" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" Oct 08 06:53:06 crc kubenswrapper[4810]: E1008 06:53:06.786476 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611\": container with ID starting with 785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611 not found: ID does not exist" containerID="785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.786530 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611"} err="failed to get container status \"785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611\": rpc error: code = NotFound desc = could not find container \"785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611\": container with ID starting with 785f34c1afdd6566a51183c720622c2d929a44a8d04bea207ea7b24be6030611 not found: ID does not exist" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.786555 4810 scope.go:117] "RemoveContainer" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" Oct 08 06:53:06 crc kubenswrapper[4810]: E1008 06:53:06.786899 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775\": container with ID starting with f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 not found: ID does not exist" containerID="f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.786922 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775"} err="failed to get container status \"f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775\": rpc error: code = NotFound desc = could not find container \"f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775\": container with ID starting with f742be9e3eb81973a065bc698509eb90800dad52fb7d890a5f5945cd7c849775 not found: ID does not exist" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.786935 4810 scope.go:117] "RemoveContainer" containerID="5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537" Oct 08 06:53:06 crc kubenswrapper[4810]: E1008 06:53:06.787367 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537\": container with ID starting with 5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537 not found: ID does not exist" containerID="5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537" Oct 08 06:53:06 crc kubenswrapper[4810]: I1008 06:53:06.787410 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537"} err="failed to get container status \"5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537\": rpc error: code = NotFound desc = could not find container \"5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537\": container with ID starting with 5496117d5ec9298a37d8b2ab727ea31675ba61d4c57589aa728e487071f31537 not found: ID does not exist" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.011418 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.148325 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.148501 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") pod \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.148577 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-cache\") pod \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.148609 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfvmk\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-kube-api-access-rfvmk\") pod \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.148629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-lock\") pod \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\" (UID: \"63aa00a9-34a1-45ce-9ada-c767d07eabb8\") " Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.149330 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-lock" (OuterVolumeSpecName: "lock") pod "63aa00a9-34a1-45ce-9ada-c767d07eabb8" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.149880 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-cache" (OuterVolumeSpecName: "cache") pod "63aa00a9-34a1-45ce-9ada-c767d07eabb8" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.154929 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "63aa00a9-34a1-45ce-9ada-c767d07eabb8" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.155363 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "swift") pod "63aa00a9-34a1-45ce-9ada-c767d07eabb8" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.155647 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-kube-api-access-rfvmk" (OuterVolumeSpecName: "kube-api-access-rfvmk") pod "63aa00a9-34a1-45ce-9ada-c767d07eabb8" (UID: "63aa00a9-34a1-45ce-9ada-c767d07eabb8"). InnerVolumeSpecName "kube-api-access-rfvmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.250757 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.250797 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.250807 4810 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-cache\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.250818 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfvmk\" (UniqueName: \"kubernetes.io/projected/63aa00a9-34a1-45ce-9ada-c767d07eabb8-kube-api-access-rfvmk\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.250828 4810 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/63aa00a9-34a1-45ce-9ada-c767d07eabb8-lock\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.267093 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.352269 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.722585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"63aa00a9-34a1-45ce-9ada-c767d07eabb8","Type":"ContainerDied","Data":"cfbfca7b2ba7b12331603d4ad82d621e062bc8d874c683ffabae0d62b492524e"} Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.722664 4810 scope.go:117] "RemoveContainer" containerID="758c08848dcf56803d3340df9307815df8fed52c268f9d576bd87dce38f15071" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.722783 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.777070 4810 scope.go:117] "RemoveContainer" containerID="b0eb359ab07ae0f5e3fe4ee340568b60987dd5232e6f155a7a05a079ef8c5e07" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.777833 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.783099 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.816887 4810 scope.go:117] "RemoveContainer" containerID="345bb23b7033db196c378bf859a6c3b70c55475aa015ae3af5ecd4503e2dcf38" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.847245 4810 scope.go:117] "RemoveContainer" containerID="600f7e0d7b7425e861bb2335da4e0b3bf0e9e3610e33eb0340d0e575eba425da" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.879464 4810 scope.go:117] "RemoveContainer" containerID="f46dd8e3225e4f46200299b022317c9b2569c98b96714c24a69495ae35120418" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.908697 4810 scope.go:117] "RemoveContainer" containerID="e5382afb7813f143406f3f3d996682bc6f5be13125adec6966b7d947409db0b5" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.942160 4810 scope.go:117] "RemoveContainer" containerID="d8f77941e09530d2318f1aa89644df82f0915aa3cfe954e93ebcfc8a0be63897" Oct 08 06:53:07 crc kubenswrapper[4810]: I1008 06:53:07.972422 4810 scope.go:117] "RemoveContainer" containerID="b1c09d7235386cf593d3456b3d746d6d1cf962a7a681c42f5eb436bbfa5c7423" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.058396 4810 scope.go:117] "RemoveContainer" containerID="aeceb614fb9899cfafb9c3fa078fce2b620e618a9756ff3d62a4a160e354a0b2" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.085471 4810 scope.go:117] "RemoveContainer" containerID="8c66da6cd63ffdac911d7e4cb28fc214ac4f2ca0696b970109e6193d96655d61" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.090487 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" path="/var/lib/kubelet/pods/63aa00a9-34a1-45ce-9ada-c767d07eabb8/volumes" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.093162 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" path="/var/lib/kubelet/pods/fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37/volumes" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.107800 4810 scope.go:117] "RemoveContainer" containerID="ce160cdae1ce3f85d7b348919fd8ec1e63ee09bafdd24ff6e4c23db7ad5a2ff1" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.152385 4810 scope.go:117] "RemoveContainer" containerID="f2a45c96c2a4d65f22eb75f3b31681ceb8d4be034c534c71d36192573c688352" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.182476 4810 scope.go:117] "RemoveContainer" containerID="cc83732eb2aa89586c96fc2f8b3a8d349fb000018c5d29cd6d05bd9035d6901c" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.212356 4810 scope.go:117] "RemoveContainer" containerID="f97fcf1457257c718d7bc50e69bb8e879b0bcdda1152aafd198dc4eacc3764c1" Oct 08 06:53:08 crc kubenswrapper[4810]: I1008 06:53:08.235370 4810 scope.go:117] "RemoveContainer" containerID="39e49e136cc29f048dd01f7cdb829f4235eb528d5956a0e907b4abc94d91b6bd" Oct 08 06:53:19 crc kubenswrapper[4810]: I1008 06:53:19.431881 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:53:19 crc kubenswrapper[4810]: I1008 06:53:19.432598 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:53:41 crc kubenswrapper[4810]: I1008 06:53:41.287296 4810 scope.go:117] "RemoveContainer" containerID="85cbea479eda95f8290d55a3c5e21c38ca10eb0dbc85ea481f1c6dcd9a95fdc7" Oct 08 06:53:41 crc kubenswrapper[4810]: I1008 06:53:41.331239 4810 scope.go:117] "RemoveContainer" containerID="dd78d50a7f1c3827da6046339729794dc53b8121adbd9c479e00e2f78cbf711a" Oct 08 06:53:49 crc kubenswrapper[4810]: I1008 06:53:49.431387 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:53:49 crc kubenswrapper[4810]: I1008 06:53:49.431943 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:53:49 crc kubenswrapper[4810]: I1008 06:53:49.432024 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:53:49 crc kubenswrapper[4810]: I1008 06:53:49.432830 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1fd9fd64ac269c2eaf81957ab40e460fd424d99aa24609f1272b49a186f84120"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:53:49 crc kubenswrapper[4810]: I1008 06:53:49.432923 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://1fd9fd64ac269c2eaf81957ab40e460fd424d99aa24609f1272b49a186f84120" gracePeriod=600 Oct 08 06:53:50 crc kubenswrapper[4810]: I1008 06:53:50.195656 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="1fd9fd64ac269c2eaf81957ab40e460fd424d99aa24609f1272b49a186f84120" exitCode=0 Oct 08 06:53:50 crc kubenswrapper[4810]: I1008 06:53:50.195741 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"1fd9fd64ac269c2eaf81957ab40e460fd424d99aa24609f1272b49a186f84120"} Oct 08 06:53:50 crc kubenswrapper[4810]: I1008 06:53:50.196351 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b"} Oct 08 06:53:50 crc kubenswrapper[4810]: I1008 06:53:50.196376 4810 scope.go:117] "RemoveContainer" containerID="fe33a8b708a76b97f4ea809aa4cdda57b3d8c666bcf720b830743227cfcf88de" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.877711 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-twzd2"] Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.878822 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5091c146-f092-4878-95a3-49cb91760d17" containerName="init" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.878837 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5091c146-f092-4878-95a3-49cb91760d17" containerName="init" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.878855 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.878861 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-server" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.878872 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879035 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879048 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879055 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879069 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server-init" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879076 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server-init" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879083 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-updater" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879089 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-updater" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879121 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879127 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879139 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73271be2-0048-432e-b389-183258582254" containerName="nova-scheduler-scheduler" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879145 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="73271be2-0048-432e-b389-183258582254" containerName="nova-scheduler-scheduler" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879152 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879158 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879167 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9895842c-0b86-4c8c-9076-3777d45634dd" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879173 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9895842c-0b86-4c8c-9076-3777d45634dd" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879186 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879191 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879202 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879208 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879215 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-notification-agent" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879224 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-notification-agent" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879230 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879236 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879250 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4deb0612-547f-4067-b95b-5794663d21aa" containerName="setup-container" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879255 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4deb0612-547f-4067-b95b-5794663d21aa" containerName="setup-container" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879261 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-updater" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879267 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-updater" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879276 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" containerName="kube-state-metrics" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879282 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" containerName="kube-state-metrics" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879288 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c1351c1-21aa-4004-a402-a603343c58d7" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879294 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c1351c1-21aa-4004-a402-a603343c58d7" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879302 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879307 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-server" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879317 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebbf3e4c-75db-4d41-8220-4b5c687d815f" containerName="nova-cell0-conductor-conductor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879323 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebbf3e4c-75db-4d41-8220-4b5c687d815f" containerName="nova-cell0-conductor-conductor" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879333 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879338 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879344 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a940c7a1-b3cc-4168-99e8-750c87845314" containerName="mysql-bootstrap" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879351 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a940c7a1-b3cc-4168-99e8-750c87845314" containerName="mysql-bootstrap" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879358 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a940c7a1-b3cc-4168-99e8-750c87845314" containerName="galera" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879363 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a940c7a1-b3cc-4168-99e8-750c87845314" containerName="galera" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879373 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="rabbitmq" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879378 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="rabbitmq" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879385 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="cinder-scheduler" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879391 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="cinder-scheduler" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879399 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879404 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-api" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879411 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879416 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879424 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-expirer" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879429 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-expirer" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879436 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879442 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-server" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879450 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" containerName="nova-cell1-conductor-conductor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879457 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" containerName="nova-cell1-conductor-conductor" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879463 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="sg-core" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879468 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="sg-core" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879475 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerName="galera" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879482 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerName="galera" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879489 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-reaper" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879495 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-reaper" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879504 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerName="mysql-bootstrap" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879509 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerName="mysql-bootstrap" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879516 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b1ed43-571a-4b70-9eae-c19a7675bc59" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879522 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b1ed43-571a-4b70-9eae-c19a7675bc59" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879533 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="rsync" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879538 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="rsync" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879547 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879552 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879559 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="proxy-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879564 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="proxy-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879574 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="probe" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879579 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="probe" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879585 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879590 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879598 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced62b16-201e-4f75-a970-20bad14e1441" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879603 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced62b16-201e-4f75-a970-20bad14e1441" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879611 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="ovsdbserver-nb" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879616 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="ovsdbserver-nb" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879624 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879630 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879638 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879643 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879650 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879655 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879665 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879671 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879679 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879685 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879693 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="ovsdbserver-sb" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879699 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="ovsdbserver-sb" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879706 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879711 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879721 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879727 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-server" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879735 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="ovn-northd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879741 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="ovn-northd" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879747 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879753 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879762 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49cb68f6-1a48-4d79-90cf-34897b19dfaa" containerName="memcached" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879768 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="49cb68f6-1a48-4d79-90cf-34897b19dfaa" containerName="memcached" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879775 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="swift-recon-cron" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879780 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="swift-recon-cron" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879791 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879797 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-api" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879805 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879811 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879818 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879824 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879831 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879836 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879844 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879849 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879855 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879860 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-api" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879870 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f5fbdd-0691-4ce8-888d-be0147974607" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879875 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f5fbdd-0691-4ce8-888d-be0147974607" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879883 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879890 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879899 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493810ad-f07f-4871-8063-5d54cd5786ea" containerName="keystone-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879906 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="493810ad-f07f-4871-8063-5d54cd5786ea" containerName="keystone-api" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879914 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-central-agent" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879921 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-central-agent" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879934 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="setup-container" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879939 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="setup-container" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879948 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879953 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879980 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.879985 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.879995 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5091c146-f092-4878-95a3-49cb91760d17" containerName="dnsmasq-dns" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880000 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5091c146-f092-4878-95a3-49cb91760d17" containerName="dnsmasq-dns" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880009 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90da3b83-b1b7-4d11-9f4d-37cf76b93edb" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880014 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="90da3b83-b1b7-4d11-9f4d-37cf76b93edb" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880023 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880029 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880036 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880041 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880051 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4deb0612-547f-4067-b95b-5794663d21aa" containerName="rabbitmq" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880057 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4deb0612-547f-4067-b95b-5794663d21aa" containerName="rabbitmq" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880065 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-metadata" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880071 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-metadata" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880077 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64350cd1-38b8-4c64-8c01-105432d1b91f" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880083 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="64350cd1-38b8-4c64-8c01-105432d1b91f" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880091 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880096 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-log" Oct 08 06:54:38 crc kubenswrapper[4810]: E1008 06:54:38.880104 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880110 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880236 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-updater" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880247 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebbf3e4c-75db-4d41-8220-4b5c687d815f" containerName="nova-cell0-conductor-conductor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880256 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880269 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880275 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880286 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880292 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880298 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8bc39bc-b2a2-4a73-a0e1-1379ecc18f7b" containerName="nova-cell1-conductor-conductor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880307 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a1b9d9-5ef5-4d61-85e2-e3c1bb7f41ce" containerName="kube-state-metrics" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880318 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880328 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="49cb68f6-1a48-4d79-90cf-34897b19dfaa" containerName="memcached" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880337 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880346 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880357 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="111360ca-3afe-4eb6-83a0-c60ca579c82b" containerName="ovn-controller" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880368 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880375 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-updater" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880381 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovs-vswitchd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880389 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="sg-core" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880399 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880408 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880414 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="016c35ce-3072-485f-86b8-bbe927f78f81" containerName="barbican-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880423 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b1ed43-571a-4b70-9eae-c19a7675bc59" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880431 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880437 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880448 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c1351c1-21aa-4004-a402-a603343c58d7" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880458 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880464 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ba387c-c5ce-4f6e-8323-7e6afce894ba" containerName="ovsdbserver-sb" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880471 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-reaper" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880481 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9895842c-0b86-4c8c-9076-3777d45634dd" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880488 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5091c146-f092-4878-95a3-49cb91760d17" containerName="dnsmasq-dns" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880494 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4deb0612-547f-4067-b95b-5794663d21aa" containerName="rabbitmq" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880504 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880514 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aabc3b24-1971-4326-aff0-12ddbf1f5ea1" containerName="barbican-keystone-listener-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880521 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880530 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="cinder-scheduler" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880539 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="swift-recon-cron" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880548 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="64350cd1-38b8-4c64-8c01-105432d1b91f" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880556 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44c96c-b96c-409c-9c9e-c049d9fe68b5" containerName="rabbitmq" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880564 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-replicator" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880571 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880577 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced62b16-201e-4f75-a970-20bad14e1441" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880608 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f5fbdd-0691-4ce8-888d-be0147974607" containerName="openstack-network-exporter" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880616 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880623 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880632 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880639 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="06237db0-424d-4bd7-ae4f-3f7af9cc4f92" containerName="proxy-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880647 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="rsync" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880655 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b86049c-4b80-4a1f-8e8b-51ad6e174fd7" containerName="galera" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880664 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="12678e95-0510-4cd6-86c8-701ea97a247e" containerName="probe" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880670 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc5c4854-7fbb-4810-9dfc-cc22d6cdbd37" containerName="ovsdb-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880678 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="container-auditor" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880688 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e33182-d302-417e-8e4e-fec4808047a4" containerName="cinder-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880695 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ca6de6-470e-4226-b67b-e58bf64415e3" containerName="glance-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880703 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd96fb6-493a-47d6-99e9-f3acfcc4c7a7" containerName="ovn-northd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880712 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="account-server" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880721 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d95dc4d-f167-494a-b09f-09b5e0cf8ab0" containerName="neutron-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880729 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8832ab0b-b264-4db9-8b70-87920c1e4826" containerName="nova-metadata-metadata" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880735 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a3bceca-c9a2-45e1-9f81-2bd9f6985c36" containerName="ovsdbserver-nb" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880745 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-central-agent" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880754 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f266b088-0be1-4d57-b8f9-c891e8ad1ed3" containerName="nova-api-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880762 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="90da3b83-b1b7-4d11-9f4d-37cf76b93edb" containerName="mariadb-account-delete" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880768 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="ceilometer-notification-agent" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880775 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a940c7a1-b3cc-4168-99e8-750c87845314" containerName="galera" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880783 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="493810ad-f07f-4871-8063-5d54cd5786ea" containerName="keystone-api" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880790 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74ea5cd-1e87-40b3-b7f4-262fdc1779d1" containerName="barbican-worker" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880797 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dced1083-d856-4f53-bb6e-19fc64ba95a5" containerName="glance-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880979 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7486bd0-d5fd-4d4f-8596-705cc22237d5" containerName="placement-log" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880988 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d58cc0-e60f-49c7-ad4c-d4db94cea14d" containerName="proxy-httpd" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.880995 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="73271be2-0048-432e-b389-183258582254" containerName="nova-scheduler-scheduler" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.881004 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aa00a9-34a1-45ce-9ada-c767d07eabb8" containerName="object-expirer" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.882655 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.889356 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-twzd2"] Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.988642 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkz46\" (UniqueName: \"kubernetes.io/projected/be72a8e9-b591-4886-abf4-3647d9b1ae80-kube-api-access-fkz46\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.989631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-catalog-content\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:38 crc kubenswrapper[4810]: I1008 06:54:38.989848 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-utilities\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.092101 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-utilities\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.092165 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkz46\" (UniqueName: \"kubernetes.io/projected/be72a8e9-b591-4886-abf4-3647d9b1ae80-kube-api-access-fkz46\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.092228 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-catalog-content\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.092644 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-utilities\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.092816 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-catalog-content\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.116856 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkz46\" (UniqueName: \"kubernetes.io/projected/be72a8e9-b591-4886-abf4-3647d9b1ae80-kube-api-access-fkz46\") pod \"redhat-operators-twzd2\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.229369 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.639218 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-twzd2"] Oct 08 06:54:39 crc kubenswrapper[4810]: W1008 06:54:39.655443 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe72a8e9_b591_4886_abf4_3647d9b1ae80.slice/crio-d9589e4f0c5d1546b03dda11ffad819cfa34bdf85dde002d62a8f140c8c22724 WatchSource:0}: Error finding container d9589e4f0c5d1546b03dda11ffad819cfa34bdf85dde002d62a8f140c8c22724: Status 404 returned error can't find the container with id d9589e4f0c5d1546b03dda11ffad819cfa34bdf85dde002d62a8f140c8c22724 Oct 08 06:54:39 crc kubenswrapper[4810]: I1008 06:54:39.694851 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twzd2" event={"ID":"be72a8e9-b591-4886-abf4-3647d9b1ae80","Type":"ContainerStarted","Data":"d9589e4f0c5d1546b03dda11ffad819cfa34bdf85dde002d62a8f140c8c22724"} Oct 08 06:54:40 crc kubenswrapper[4810]: I1008 06:54:40.709042 4810 generic.go:334] "Generic (PLEG): container finished" podID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerID="dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69" exitCode=0 Oct 08 06:54:40 crc kubenswrapper[4810]: I1008 06:54:40.709111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twzd2" event={"ID":"be72a8e9-b591-4886-abf4-3647d9b1ae80","Type":"ContainerDied","Data":"dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69"} Oct 08 06:54:40 crc kubenswrapper[4810]: I1008 06:54:40.711545 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.721422 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twzd2" event={"ID":"be72a8e9-b591-4886-abf4-3647d9b1ae80","Type":"ContainerStarted","Data":"b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674"} Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.775047 4810 scope.go:117] "RemoveContainer" containerID="21bf0b67108876603e6a3ca6e2697321f8c85f0591ff8f0d50ebf6f57df7b7b3" Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.805623 4810 scope.go:117] "RemoveContainer" containerID="cb62f083c1a6339ca4b73ea3bc19a767cc24f149de31aa454da4c2003b472a7c" Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.865065 4810 scope.go:117] "RemoveContainer" containerID="e5867c8c5052a231dc51cd45cde9fddb5279855338c65d3dc46514e20fd68fcf" Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.888945 4810 scope.go:117] "RemoveContainer" containerID="5a780ab4f9043af07e2e7ba563e150db2ea7db7ad6e6b93d0b0d6250802624cd" Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.914372 4810 scope.go:117] "RemoveContainer" containerID="f60d12361f4468b41bc7cb57d821f3097194d727ab08709eb1949abd417c69a2" Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.938466 4810 scope.go:117] "RemoveContainer" containerID="304adddb1efc0687866b5c780e8dddddb285afdc0117a61a17f3af47c6303e9f" Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.961568 4810 scope.go:117] "RemoveContainer" containerID="eacb2363992ea55cecd62e4498198b8367c189671343114cf21204d0ef93a535" Oct 08 06:54:41 crc kubenswrapper[4810]: I1008 06:54:41.983203 4810 scope.go:117] "RemoveContainer" containerID="80569e9d82763f7df4441c225a43c04f1089ccfd570f2a0b75c3f7ec147cdb62" Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.008505 4810 scope.go:117] "RemoveContainer" containerID="7456a6a32595ebd5c9681d2ead99703593de54cb42b0b538d1d12ebd1daeb457" Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.090731 4810 scope.go:117] "RemoveContainer" containerID="467ac7e064869303b6ece8fd50d804b1d6f99a7fadaee0b3b8c2839cf2711511" Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.118419 4810 scope.go:117] "RemoveContainer" containerID="3a567dca81c9cd9b5bcaec9dfaa1354ae1f435226e25175f097dce75fffa0d2e" Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.183428 4810 scope.go:117] "RemoveContainer" containerID="95cc241da1c907b8dff8d4520a8e079f4946bdfb0285b71777b56e61b76c747e" Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.200212 4810 scope.go:117] "RemoveContainer" containerID="28edf5eb737efa28788525bfeb7f9c64a807fad3bc80060ae62756f6fad80ae2" Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.226200 4810 scope.go:117] "RemoveContainer" containerID="9c36ae631cf8df1c181b11732cc17ae7be213aa863fab14cd460920c2dd2e53d" Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.739212 4810 generic.go:334] "Generic (PLEG): container finished" podID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerID="b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674" exitCode=0 Oct 08 06:54:42 crc kubenswrapper[4810]: I1008 06:54:42.739267 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twzd2" event={"ID":"be72a8e9-b591-4886-abf4-3647d9b1ae80","Type":"ContainerDied","Data":"b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674"} Oct 08 06:54:43 crc kubenswrapper[4810]: I1008 06:54:43.752642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twzd2" event={"ID":"be72a8e9-b591-4886-abf4-3647d9b1ae80","Type":"ContainerStarted","Data":"ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003"} Oct 08 06:54:43 crc kubenswrapper[4810]: I1008 06:54:43.778948 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-twzd2" podStartSLOduration=3.329864842 podStartE2EDuration="5.77892832s" podCreationTimestamp="2025-10-08 06:54:38 +0000 UTC" firstStartedPulling="2025-10-08 06:54:40.71086381 +0000 UTC m=+1383.345303560" lastFinishedPulling="2025-10-08 06:54:43.159927288 +0000 UTC m=+1385.794367038" observedRunningTime="2025-10-08 06:54:43.776512114 +0000 UTC m=+1386.410951874" watchObservedRunningTime="2025-10-08 06:54:43.77892832 +0000 UTC m=+1386.413368070" Oct 08 06:54:49 crc kubenswrapper[4810]: I1008 06:54:49.230255 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:49 crc kubenswrapper[4810]: I1008 06:54:49.230696 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:49 crc kubenswrapper[4810]: I1008 06:54:49.303949 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:49 crc kubenswrapper[4810]: I1008 06:54:49.874216 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:49 crc kubenswrapper[4810]: I1008 06:54:49.926350 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-twzd2"] Oct 08 06:54:51 crc kubenswrapper[4810]: I1008 06:54:51.839190 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-twzd2" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="registry-server" containerID="cri-o://ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003" gracePeriod=2 Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.269208 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.381680 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-utilities\") pod \"be72a8e9-b591-4886-abf4-3647d9b1ae80\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.381734 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkz46\" (UniqueName: \"kubernetes.io/projected/be72a8e9-b591-4886-abf4-3647d9b1ae80-kube-api-access-fkz46\") pod \"be72a8e9-b591-4886-abf4-3647d9b1ae80\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.381770 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-catalog-content\") pod \"be72a8e9-b591-4886-abf4-3647d9b1ae80\" (UID: \"be72a8e9-b591-4886-abf4-3647d9b1ae80\") " Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.383411 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-utilities" (OuterVolumeSpecName: "utilities") pod "be72a8e9-b591-4886-abf4-3647d9b1ae80" (UID: "be72a8e9-b591-4886-abf4-3647d9b1ae80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.389333 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be72a8e9-b591-4886-abf4-3647d9b1ae80-kube-api-access-fkz46" (OuterVolumeSpecName: "kube-api-access-fkz46") pod "be72a8e9-b591-4886-abf4-3647d9b1ae80" (UID: "be72a8e9-b591-4886-abf4-3647d9b1ae80"). InnerVolumeSpecName "kube-api-access-fkz46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.483520 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkz46\" (UniqueName: \"kubernetes.io/projected/be72a8e9-b591-4886-abf4-3647d9b1ae80-kube-api-access-fkz46\") on node \"crc\" DevicePath \"\"" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.483579 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.854379 4810 generic.go:334] "Generic (PLEG): container finished" podID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerID="ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003" exitCode=0 Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.854492 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twzd2" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.854542 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twzd2" event={"ID":"be72a8e9-b591-4886-abf4-3647d9b1ae80","Type":"ContainerDied","Data":"ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003"} Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.855748 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twzd2" event={"ID":"be72a8e9-b591-4886-abf4-3647d9b1ae80","Type":"ContainerDied","Data":"d9589e4f0c5d1546b03dda11ffad819cfa34bdf85dde002d62a8f140c8c22724"} Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.855817 4810 scope.go:117] "RemoveContainer" containerID="ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.892442 4810 scope.go:117] "RemoveContainer" containerID="b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.918591 4810 scope.go:117] "RemoveContainer" containerID="dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.952926 4810 scope.go:117] "RemoveContainer" containerID="ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003" Oct 08 06:54:52 crc kubenswrapper[4810]: E1008 06:54:52.953608 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003\": container with ID starting with ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003 not found: ID does not exist" containerID="ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.953660 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003"} err="failed to get container status \"ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003\": rpc error: code = NotFound desc = could not find container \"ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003\": container with ID starting with ec37358b2309df93aab9a740e94b6c0010ddf4dcbdb48455c783c8a6c42b9003 not found: ID does not exist" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.953695 4810 scope.go:117] "RemoveContainer" containerID="b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674" Oct 08 06:54:52 crc kubenswrapper[4810]: E1008 06:54:52.954304 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674\": container with ID starting with b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674 not found: ID does not exist" containerID="b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.954499 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674"} err="failed to get container status \"b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674\": rpc error: code = NotFound desc = could not find container \"b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674\": container with ID starting with b548b0c0bc60e66e8ac9b3885c8155de40349e4b2376e4d446902d8954326674 not found: ID does not exist" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.954560 4810 scope.go:117] "RemoveContainer" containerID="dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69" Oct 08 06:54:52 crc kubenswrapper[4810]: E1008 06:54:52.955401 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69\": container with ID starting with dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69 not found: ID does not exist" containerID="dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69" Oct 08 06:54:52 crc kubenswrapper[4810]: I1008 06:54:52.955454 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69"} err="failed to get container status \"dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69\": rpc error: code = NotFound desc = could not find container \"dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69\": container with ID starting with dd9c0745da6d29f87bd9da3e23076812bfaff28f870ec1d5943716a39ce8ff69 not found: ID does not exist" Oct 08 06:54:53 crc kubenswrapper[4810]: I1008 06:54:53.166712 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be72a8e9-b591-4886-abf4-3647d9b1ae80" (UID: "be72a8e9-b591-4886-abf4-3647d9b1ae80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:54:53 crc kubenswrapper[4810]: I1008 06:54:53.196926 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be72a8e9-b591-4886-abf4-3647d9b1ae80-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:54:53 crc kubenswrapper[4810]: I1008 06:54:53.514919 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-twzd2"] Oct 08 06:54:53 crc kubenswrapper[4810]: I1008 06:54:53.526746 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-twzd2"] Oct 08 06:54:54 crc kubenswrapper[4810]: I1008 06:54:54.092173 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" path="/var/lib/kubelet/pods/be72a8e9-b591-4886-abf4-3647d9b1ae80/volumes" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.161569 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9pghs"] Oct 08 06:55:42 crc kubenswrapper[4810]: E1008 06:55:42.165405 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="extract-content" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.165719 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="extract-content" Oct 08 06:55:42 crc kubenswrapper[4810]: E1008 06:55:42.166046 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="registry-server" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.166266 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="registry-server" Oct 08 06:55:42 crc kubenswrapper[4810]: E1008 06:55:42.166517 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="extract-utilities" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.166707 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="extract-utilities" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.167384 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="be72a8e9-b591-4886-abf4-3647d9b1ae80" containerName="registry-server" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.170452 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.178403 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9pghs"] Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.295686 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxhx4\" (UniqueName: \"kubernetes.io/projected/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-kube-api-access-gxhx4\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.295789 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-utilities\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.295851 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-catalog-content\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.397329 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-catalog-content\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.397702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxhx4\" (UniqueName: \"kubernetes.io/projected/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-kube-api-access-gxhx4\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.397930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-utilities\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.399005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-utilities\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.401492 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-catalog-content\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.418197 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxhx4\" (UniqueName: \"kubernetes.io/projected/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-kube-api-access-gxhx4\") pod \"community-operators-9pghs\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.460148 4810 scope.go:117] "RemoveContainer" containerID="6a3956be0e123bf75016806de880277e6c672f2d8443f2f7b9d2fbda532d3217" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.486626 4810 scope.go:117] "RemoveContainer" containerID="1496b7ed265837ae3927cb59497e3d39ba71ca3553ca631f821eef258bd3755f" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.504540 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.548388 4810 scope.go:117] "RemoveContainer" containerID="3556449533e62b8a1b045f6331743feae434169e2723d25a19669ad4c27722e9" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.610235 4810 scope.go:117] "RemoveContainer" containerID="76a5e6948d4daf32096a3e2e398fd4087c887c28e01b46a3346cda1ecc593c0c" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.651483 4810 scope.go:117] "RemoveContainer" containerID="e2fd952082a2646936a99d8f5498c1dc178476cd9cb299a67544c11331778da9" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.685080 4810 scope.go:117] "RemoveContainer" containerID="f1481a876ddf5f2a5996c05b5e44aa7cfa1cf846a5a9246337a1abce43f43d18" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.744856 4810 scope.go:117] "RemoveContainer" containerID="e3d36cfbaa8984ab3a07deadd18a85cb569ec7b4cff4e318a3ad985e94a1df6a" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.795778 4810 scope.go:117] "RemoveContainer" containerID="7d46128d5daaaf86caffd08803a49b7c5f8c2451b094dba375d4a115f05b3abd" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.823840 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9pghs"] Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.843286 4810 scope.go:117] "RemoveContainer" containerID="b526413b9a8c8e6cae317ae5e79f9dced75f63320d48b38b7a09dab4a2bdcb30" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.860150 4810 scope.go:117] "RemoveContainer" containerID="17af0d67e7c101fc6c86a1b89cffcf25c43ff30824d53b0441e8c67f457254be" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.886074 4810 scope.go:117] "RemoveContainer" containerID="f651cc8c1eb5d6258b63b8e4bc2df041f992fce6e5d8e0d734458bee6b085f54" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.900497 4810 scope.go:117] "RemoveContainer" containerID="66a0ef3b680887a96f474c84ba92ef0028c19cf0633bc4c7f831bb1f67bcc308" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.931199 4810 scope.go:117] "RemoveContainer" containerID="8886bbf301fede3b944ade8f43dbcf03b0120838fdcffa8ac339943a18759597" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.965680 4810 scope.go:117] "RemoveContainer" containerID="8b1378e928905e98ed227c844df3846d51fa404c6a1cb6f7629ca5ddef749cff" Oct 08 06:55:42 crc kubenswrapper[4810]: I1008 06:55:42.979885 4810 scope.go:117] "RemoveContainer" containerID="541ed13448c8e2c968839050706935ecc2bebbb1bbc55d3025a7f620c1798afd" Oct 08 06:55:43 crc kubenswrapper[4810]: I1008 06:55:43.388994 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pghs" event={"ID":"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411","Type":"ContainerDied","Data":"18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091"} Oct 08 06:55:43 crc kubenswrapper[4810]: I1008 06:55:43.388957 4810 generic.go:334] "Generic (PLEG): container finished" podID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerID="18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091" exitCode=0 Oct 08 06:55:43 crc kubenswrapper[4810]: I1008 06:55:43.389056 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pghs" event={"ID":"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411","Type":"ContainerStarted","Data":"60fcaa45737800117b9e57f3d02d12bd4a8638ce6ad77eb05addd6616554bdec"} Oct 08 06:55:44 crc kubenswrapper[4810]: I1008 06:55:44.398392 4810 generic.go:334] "Generic (PLEG): container finished" podID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerID="3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a" exitCode=0 Oct 08 06:55:44 crc kubenswrapper[4810]: I1008 06:55:44.398440 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pghs" event={"ID":"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411","Type":"ContainerDied","Data":"3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a"} Oct 08 06:55:45 crc kubenswrapper[4810]: I1008 06:55:45.410053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pghs" event={"ID":"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411","Type":"ContainerStarted","Data":"1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e"} Oct 08 06:55:45 crc kubenswrapper[4810]: I1008 06:55:45.438997 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9pghs" podStartSLOduration=1.980901523 podStartE2EDuration="3.438934008s" podCreationTimestamp="2025-10-08 06:55:42 +0000 UTC" firstStartedPulling="2025-10-08 06:55:43.390615406 +0000 UTC m=+1446.025055146" lastFinishedPulling="2025-10-08 06:55:44.848647891 +0000 UTC m=+1447.483087631" observedRunningTime="2025-10-08 06:55:45.43683455 +0000 UTC m=+1448.071274300" watchObservedRunningTime="2025-10-08 06:55:45.438934008 +0000 UTC m=+1448.073373748" Oct 08 06:55:49 crc kubenswrapper[4810]: I1008 06:55:49.431462 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:55:49 crc kubenswrapper[4810]: I1008 06:55:49.432065 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:55:52 crc kubenswrapper[4810]: I1008 06:55:52.505472 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:52 crc kubenswrapper[4810]: I1008 06:55:52.505787 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:52 crc kubenswrapper[4810]: I1008 06:55:52.569371 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:53 crc kubenswrapper[4810]: I1008 06:55:53.548416 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:53 crc kubenswrapper[4810]: I1008 06:55:53.595990 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9pghs"] Oct 08 06:55:55 crc kubenswrapper[4810]: I1008 06:55:55.499319 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9pghs" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="registry-server" containerID="cri-o://1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e" gracePeriod=2 Oct 08 06:55:55 crc kubenswrapper[4810]: I1008 06:55:55.895037 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.005884 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-utilities\") pod \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.006370 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-catalog-content\") pod \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.006494 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxhx4\" (UniqueName: \"kubernetes.io/projected/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-kube-api-access-gxhx4\") pod \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\" (UID: \"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411\") " Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.007226 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-utilities" (OuterVolumeSpecName: "utilities") pod "e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" (UID: "e8bcbd3e-510a-4c7b-aae2-c0a9cef11411"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.013134 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-kube-api-access-gxhx4" (OuterVolumeSpecName: "kube-api-access-gxhx4") pod "e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" (UID: "e8bcbd3e-510a-4c7b-aae2-c0a9cef11411"). InnerVolumeSpecName "kube-api-access-gxhx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.089739 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" (UID: "e8bcbd3e-510a-4c7b-aae2-c0a9cef11411"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.108766 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.108819 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.108840 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxhx4\" (UniqueName: \"kubernetes.io/projected/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411-kube-api-access-gxhx4\") on node \"crc\" DevicePath \"\"" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.512268 4810 generic.go:334] "Generic (PLEG): container finished" podID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerID="1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e" exitCode=0 Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.512329 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pghs" event={"ID":"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411","Type":"ContainerDied","Data":"1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e"} Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.512372 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pghs" event={"ID":"e8bcbd3e-510a-4c7b-aae2-c0a9cef11411","Type":"ContainerDied","Data":"60fcaa45737800117b9e57f3d02d12bd4a8638ce6ad77eb05addd6616554bdec"} Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.512401 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pghs" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.512416 4810 scope.go:117] "RemoveContainer" containerID="1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.538708 4810 scope.go:117] "RemoveContainer" containerID="3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.544412 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9pghs"] Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.549934 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9pghs"] Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.567417 4810 scope.go:117] "RemoveContainer" containerID="18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.593139 4810 scope.go:117] "RemoveContainer" containerID="1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e" Oct 08 06:55:56 crc kubenswrapper[4810]: E1008 06:55:56.593837 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e\": container with ID starting with 1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e not found: ID does not exist" containerID="1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.593890 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e"} err="failed to get container status \"1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e\": rpc error: code = NotFound desc = could not find container \"1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e\": container with ID starting with 1c7dc2ea4389386212ec643ccc1e6ec7fded80f881288c8d824b3349e70f0c1e not found: ID does not exist" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.593922 4810 scope.go:117] "RemoveContainer" containerID="3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a" Oct 08 06:55:56 crc kubenswrapper[4810]: E1008 06:55:56.594336 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a\": container with ID starting with 3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a not found: ID does not exist" containerID="3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.594443 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a"} err="failed to get container status \"3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a\": rpc error: code = NotFound desc = could not find container \"3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a\": container with ID starting with 3b781bee8068b464779869f78198228491c997de9de907c8e1c9f963a05c133a not found: ID does not exist" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.594554 4810 scope.go:117] "RemoveContainer" containerID="18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091" Oct 08 06:55:56 crc kubenswrapper[4810]: E1008 06:55:56.594978 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091\": container with ID starting with 18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091 not found: ID does not exist" containerID="18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091" Oct 08 06:55:56 crc kubenswrapper[4810]: I1008 06:55:56.595092 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091"} err="failed to get container status \"18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091\": rpc error: code = NotFound desc = could not find container \"18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091\": container with ID starting with 18f3ef3744e46340eb42afdd7bd56ecc30c74d5a51fa8b274769557645d61091 not found: ID does not exist" Oct 08 06:55:58 crc kubenswrapper[4810]: I1008 06:55:58.092731 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" path="/var/lib/kubelet/pods/e8bcbd3e-510a-4c7b-aae2-c0a9cef11411/volumes" Oct 08 06:56:19 crc kubenswrapper[4810]: I1008 06:56:19.432073 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:56:19 crc kubenswrapper[4810]: I1008 06:56:19.432713 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.152122 4810 scope.go:117] "RemoveContainer" containerID="81749551b081df35eaa781dfa19fa8b4dcb1258c7b85162b7bf3626674f4ace6" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.184926 4810 scope.go:117] "RemoveContainer" containerID="f6b4a8db2b3498b1430bdc164af3693743a79260b3c2097cdb4bee62cab40e39" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.214861 4810 scope.go:117] "RemoveContainer" containerID="811375e4fddcc6b30fb161f68a76625844e575fef371e71934b1514a60a8c6a3" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.244275 4810 scope.go:117] "RemoveContainer" containerID="1750bec16a054a265008e3c25946acd7c889394cd5239ce5e9246677d0be2833" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.267234 4810 scope.go:117] "RemoveContainer" containerID="8d3bf6f25cf1c45bf48dbb9c1a31f7c04f185f3159a9935da5533577297101f4" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.305653 4810 scope.go:117] "RemoveContainer" containerID="457d3161b5820d3665e9f9681e0892c95a60e78b6401eaba1569cde3e2859c2d" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.328549 4810 scope.go:117] "RemoveContainer" containerID="47e85962732a13a5f5460d0367ec3af0533b5a5617841d2c9fda2a5452a2fbdb" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.373294 4810 scope.go:117] "RemoveContainer" containerID="811cbf9580f361194bc54a2a641dd527653e175eccb2b963cfa94e4f09a79794" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.397934 4810 scope.go:117] "RemoveContainer" containerID="ed04a8db57a733ab9675e113837de747b29cbd4384df9603ebc90167bf592eca" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.417529 4810 scope.go:117] "RemoveContainer" containerID="af751edb5dd316cd9dd468c60b6235482bce1444060e49de080862ecf361ab3f" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.443510 4810 scope.go:117] "RemoveContainer" containerID="949b16b370e9288b713511d19a1e607e09e78c34f767d6c24c5b4e75b772a646" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.471729 4810 scope.go:117] "RemoveContainer" containerID="63b11541d3ac4eab5d5f382a400b96fb2be25d39229d2eacfc7b26bafe86a200" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.493389 4810 scope.go:117] "RemoveContainer" containerID="849c967153d2f165920e9dfd4ff8e9b6bd05ad3f465bda0698dd4e80edcf02a0" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.508450 4810 scope.go:117] "RemoveContainer" containerID="9241825651b89e0de120cb13a697f373ba57f94026cdad88f2152713fec8f376" Oct 08 06:56:43 crc kubenswrapper[4810]: I1008 06:56:43.542845 4810 scope.go:117] "RemoveContainer" containerID="193dc08e9cf522234431a6f7ee7bb9230f9702c5ab78ea227592b5701064d56e" Oct 08 06:56:49 crc kubenswrapper[4810]: I1008 06:56:49.431792 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 06:56:49 crc kubenswrapper[4810]: I1008 06:56:49.432431 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 06:56:49 crc kubenswrapper[4810]: I1008 06:56:49.432484 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 06:56:49 crc kubenswrapper[4810]: I1008 06:56:49.433212 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 06:56:49 crc kubenswrapper[4810]: I1008 06:56:49.433283 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" gracePeriod=600 Oct 08 06:56:49 crc kubenswrapper[4810]: E1008 06:56:49.562253 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:56:50 crc kubenswrapper[4810]: I1008 06:56:50.004286 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" exitCode=0 Oct 08 06:56:50 crc kubenswrapper[4810]: I1008 06:56:50.004345 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b"} Oct 08 06:56:50 crc kubenswrapper[4810]: I1008 06:56:50.004399 4810 scope.go:117] "RemoveContainer" containerID="1fd9fd64ac269c2eaf81957ab40e460fd424d99aa24609f1272b49a186f84120" Oct 08 06:56:50 crc kubenswrapper[4810]: I1008 06:56:50.005011 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:56:50 crc kubenswrapper[4810]: E1008 06:56:50.005369 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:57:02 crc kubenswrapper[4810]: I1008 06:57:02.073151 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:57:02 crc kubenswrapper[4810]: E1008 06:57:02.073869 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:57:13 crc kubenswrapper[4810]: I1008 06:57:13.074052 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:57:13 crc kubenswrapper[4810]: E1008 06:57:13.075210 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:57:28 crc kubenswrapper[4810]: I1008 06:57:28.082848 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:57:28 crc kubenswrapper[4810]: E1008 06:57:28.083690 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:57:41 crc kubenswrapper[4810]: I1008 06:57:41.073811 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:57:41 crc kubenswrapper[4810]: E1008 06:57:41.075168 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:57:43 crc kubenswrapper[4810]: I1008 06:57:43.716724 4810 scope.go:117] "RemoveContainer" containerID="623a3988e5bd08655720c75adbeab397cb0959190125b45cf629d8f3da4bdd2d" Oct 08 06:57:43 crc kubenswrapper[4810]: I1008 06:57:43.776908 4810 scope.go:117] "RemoveContainer" containerID="8a8635f25b51d99c85a053c1c6540e7eace15f2f55710edcd7fbf48a404e5395" Oct 08 06:57:43 crc kubenswrapper[4810]: I1008 06:57:43.805233 4810 scope.go:117] "RemoveContainer" containerID="b1c8620d6969a132a8688a38415693d0c85acae6386ee9cee039b3d30043d082" Oct 08 06:57:43 crc kubenswrapper[4810]: I1008 06:57:43.829776 4810 scope.go:117] "RemoveContainer" containerID="0b46a15f42225c1b9e736605e35ee6202b8d7630e3f3a9bbd912d262bd9c8ffe" Oct 08 06:57:43 crc kubenswrapper[4810]: I1008 06:57:43.856357 4810 scope.go:117] "RemoveContainer" containerID="21055dd63bbf423089e26821be7adc73c66cfc6acb4e71a03dd594f0a5233b3c" Oct 08 06:57:54 crc kubenswrapper[4810]: I1008 06:57:54.073742 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:57:54 crc kubenswrapper[4810]: E1008 06:57:54.075392 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:58:09 crc kubenswrapper[4810]: I1008 06:58:09.073096 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:58:09 crc kubenswrapper[4810]: E1008 06:58:09.073733 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:58:24 crc kubenswrapper[4810]: I1008 06:58:24.073676 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:58:24 crc kubenswrapper[4810]: E1008 06:58:24.074606 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:58:38 crc kubenswrapper[4810]: I1008 06:58:38.085783 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:58:38 crc kubenswrapper[4810]: E1008 06:58:38.092311 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:58:43 crc kubenswrapper[4810]: I1008 06:58:43.965618 4810 scope.go:117] "RemoveContainer" containerID="f1a4cf05a0397eeca7071d648a9bc8e63df17b35019fb39d5e0f9f897caf8b25" Oct 08 06:58:44 crc kubenswrapper[4810]: I1008 06:58:44.000913 4810 scope.go:117] "RemoveContainer" containerID="813c14b28553f034c580a431b8bf0d538af3dcb23cca601feb8f3060311ea67a" Oct 08 06:58:44 crc kubenswrapper[4810]: I1008 06:58:44.051943 4810 scope.go:117] "RemoveContainer" containerID="4be5952aa5c9dbf1e53e106b2255b1b5a179835e7c496e154ac96e8f874eeb8c" Oct 08 06:58:44 crc kubenswrapper[4810]: I1008 06:58:44.085411 4810 scope.go:117] "RemoveContainer" containerID="2a383918c50471b7a1bb4a62090441f5f687cf37ba900114f603b1c349f14dcb" Oct 08 06:58:50 crc kubenswrapper[4810]: I1008 06:58:50.073945 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:58:50 crc kubenswrapper[4810]: E1008 06:58:50.074592 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:59:05 crc kubenswrapper[4810]: I1008 06:59:05.073680 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:59:05 crc kubenswrapper[4810]: E1008 06:59:05.074344 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:59:17 crc kubenswrapper[4810]: I1008 06:59:17.073147 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:59:17 crc kubenswrapper[4810]: E1008 06:59:17.074115 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:59:32 crc kubenswrapper[4810]: I1008 06:59:32.073916 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:59:32 crc kubenswrapper[4810]: E1008 06:59:32.078324 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 06:59:46 crc kubenswrapper[4810]: I1008 06:59:46.083890 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 06:59:46 crc kubenswrapper[4810]: E1008 06:59:46.085316 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.074294 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:00:00 crc kubenswrapper[4810]: E1008 07:00:00.075552 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.163848 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb"] Oct 08 07:00:00 crc kubenswrapper[4810]: E1008 07:00:00.164741 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="extract-utilities" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.164815 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="extract-utilities" Oct 08 07:00:00 crc kubenswrapper[4810]: E1008 07:00:00.164845 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="registry-server" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.164856 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="registry-server" Oct 08 07:00:00 crc kubenswrapper[4810]: E1008 07:00:00.164936 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="extract-content" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.165052 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="extract-content" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.165499 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8bcbd3e-510a-4c7b-aae2-c0a9cef11411" containerName="registry-server" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.166340 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.168503 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.168634 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.170304 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb"] Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.232594 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-secret-volume\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.232941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nt69\" (UniqueName: \"kubernetes.io/projected/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-kube-api-access-7nt69\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.233003 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-config-volume\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.334415 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nt69\" (UniqueName: \"kubernetes.io/projected/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-kube-api-access-7nt69\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.334487 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-config-volume\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.334534 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-secret-volume\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.335707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-config-volume\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.341037 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-secret-volume\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.349336 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nt69\" (UniqueName: \"kubernetes.io/projected/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-kube-api-access-7nt69\") pod \"collect-profiles-29331780-rrdcb\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.495348 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:00 crc kubenswrapper[4810]: I1008 07:00:00.943811 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb"] Oct 08 07:00:01 crc kubenswrapper[4810]: I1008 07:00:01.681930 4810 generic.go:334] "Generic (PLEG): container finished" podID="61356cd7-4dcd-4e5d-b1dd-8501bdcc2443" containerID="473c46e220c30cf9c57c072a65bf9dce7cd288e049bfa79678b5896f4501e93d" exitCode=0 Oct 08 07:00:01 crc kubenswrapper[4810]: I1008 07:00:01.682052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" event={"ID":"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443","Type":"ContainerDied","Data":"473c46e220c30cf9c57c072a65bf9dce7cd288e049bfa79678b5896f4501e93d"} Oct 08 07:00:01 crc kubenswrapper[4810]: I1008 07:00:01.682419 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" event={"ID":"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443","Type":"ContainerStarted","Data":"acd7c52c858652af0219ebf7358ec76608906941b81a6a1895270fab8ee7fb97"} Oct 08 07:00:02 crc kubenswrapper[4810]: I1008 07:00:02.990273 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.176269 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nt69\" (UniqueName: \"kubernetes.io/projected/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-kube-api-access-7nt69\") pod \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.176375 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-config-volume\") pod \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.176523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-secret-volume\") pod \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\" (UID: \"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443\") " Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.177671 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-config-volume" (OuterVolumeSpecName: "config-volume") pod "61356cd7-4dcd-4e5d-b1dd-8501bdcc2443" (UID: "61356cd7-4dcd-4e5d-b1dd-8501bdcc2443"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.183703 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "61356cd7-4dcd-4e5d-b1dd-8501bdcc2443" (UID: "61356cd7-4dcd-4e5d-b1dd-8501bdcc2443"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.185745 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-kube-api-access-7nt69" (OuterVolumeSpecName: "kube-api-access-7nt69") pod "61356cd7-4dcd-4e5d-b1dd-8501bdcc2443" (UID: "61356cd7-4dcd-4e5d-b1dd-8501bdcc2443"). InnerVolumeSpecName "kube-api-access-7nt69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.277874 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nt69\" (UniqueName: \"kubernetes.io/projected/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-kube-api-access-7nt69\") on node \"crc\" DevicePath \"\"" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.277910 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.277920 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/61356cd7-4dcd-4e5d-b1dd-8501bdcc2443-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.703426 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" event={"ID":"61356cd7-4dcd-4e5d-b1dd-8501bdcc2443","Type":"ContainerDied","Data":"acd7c52c858652af0219ebf7358ec76608906941b81a6a1895270fab8ee7fb97"} Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.703956 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acd7c52c858652af0219ebf7358ec76608906941b81a6a1895270fab8ee7fb97" Oct 08 07:00:03 crc kubenswrapper[4810]: I1008 07:00:03.703568 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331780-rrdcb" Oct 08 07:00:14 crc kubenswrapper[4810]: I1008 07:00:14.073856 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:00:14 crc kubenswrapper[4810]: E1008 07:00:14.075885 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:00:25 crc kubenswrapper[4810]: I1008 07:00:25.074955 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:00:25 crc kubenswrapper[4810]: E1008 07:00:25.075748 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:00:37 crc kubenswrapper[4810]: I1008 07:00:37.073309 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:00:37 crc kubenswrapper[4810]: E1008 07:00:37.074073 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:00:50 crc kubenswrapper[4810]: I1008 07:00:50.073500 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:00:50 crc kubenswrapper[4810]: E1008 07:00:50.074602 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:01:04 crc kubenswrapper[4810]: I1008 07:01:04.074199 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:01:04 crc kubenswrapper[4810]: E1008 07:01:04.075757 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:01:16 crc kubenswrapper[4810]: I1008 07:01:16.074126 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:01:16 crc kubenswrapper[4810]: E1008 07:01:16.074894 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:01:28 crc kubenswrapper[4810]: I1008 07:01:28.078911 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:01:28 crc kubenswrapper[4810]: E1008 07:01:28.080214 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:01:39 crc kubenswrapper[4810]: I1008 07:01:39.073356 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:01:39 crc kubenswrapper[4810]: E1008 07:01:39.074322 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7vlk5_openshift-machine-config-operator(2b5ec8eb-9a64-46ae-b881-535323e3b686)\"" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" Oct 08 07:01:52 crc kubenswrapper[4810]: I1008 07:01:52.073894 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:01:52 crc kubenswrapper[4810]: I1008 07:01:52.615177 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"d02be0e8f536b325c81ec0c82d72faaf3765230f22216c32727e84a1825af9fe"} Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.564264 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lt8wj"] Oct 08 07:01:55 crc kubenswrapper[4810]: E1008 07:01:55.565595 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61356cd7-4dcd-4e5d-b1dd-8501bdcc2443" containerName="collect-profiles" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.565619 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="61356cd7-4dcd-4e5d-b1dd-8501bdcc2443" containerName="collect-profiles" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.565882 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="61356cd7-4dcd-4e5d-b1dd-8501bdcc2443" containerName="collect-profiles" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.573460 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.576406 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lt8wj"] Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.693625 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vvh8\" (UniqueName: \"kubernetes.io/projected/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-kube-api-access-5vvh8\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.693742 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-catalog-content\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.693773 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-utilities\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.794662 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-catalog-content\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.794708 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-utilities\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.794757 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vvh8\" (UniqueName: \"kubernetes.io/projected/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-kube-api-access-5vvh8\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.795227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-catalog-content\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.795483 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-utilities\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.816350 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vvh8\" (UniqueName: \"kubernetes.io/projected/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-kube-api-access-5vvh8\") pod \"certified-operators-lt8wj\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:55 crc kubenswrapper[4810]: I1008 07:01:55.906928 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:01:56 crc kubenswrapper[4810]: I1008 07:01:56.384027 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lt8wj"] Oct 08 07:01:56 crc kubenswrapper[4810]: I1008 07:01:56.640777 4810 generic.go:334] "Generic (PLEG): container finished" podID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerID="65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e" exitCode=0 Oct 08 07:01:56 crc kubenswrapper[4810]: I1008 07:01:56.640820 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lt8wj" event={"ID":"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a","Type":"ContainerDied","Data":"65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e"} Oct 08 07:01:56 crc kubenswrapper[4810]: I1008 07:01:56.640847 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lt8wj" event={"ID":"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a","Type":"ContainerStarted","Data":"e504ae8dd91e9ffd2e41674d2fbff9a81913dba2e6b55ce0dfcbccafb6712650"} Oct 08 07:01:56 crc kubenswrapper[4810]: I1008 07:01:56.642552 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.648377 4810 generic.go:334] "Generic (PLEG): container finished" podID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerID="bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb" exitCode=0 Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.648452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lt8wj" event={"ID":"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a","Type":"ContainerDied","Data":"bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb"} Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.749372 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q6n45"] Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.751084 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.765647 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6n45"] Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.821279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-catalog-content\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.821335 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6vlt\" (UniqueName: \"kubernetes.io/projected/4694fb39-e621-47f0-9541-423a2f6f9e4b-kube-api-access-n6vlt\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.821403 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-utilities\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.924206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-catalog-content\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.924298 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6vlt\" (UniqueName: \"kubernetes.io/projected/4694fb39-e621-47f0-9541-423a2f6f9e4b-kube-api-access-n6vlt\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.924342 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-utilities\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.924801 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-catalog-content\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.925588 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-utilities\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:57 crc kubenswrapper[4810]: I1008 07:01:57.952401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6vlt\" (UniqueName: \"kubernetes.io/projected/4694fb39-e621-47f0-9541-423a2f6f9e4b-kube-api-access-n6vlt\") pod \"redhat-marketplace-q6n45\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:58 crc kubenswrapper[4810]: I1008 07:01:58.101744 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:01:58 crc kubenswrapper[4810]: W1008 07:01:58.361783 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4694fb39_e621_47f0_9541_423a2f6f9e4b.slice/crio-d785ce62e296d0520fedbfbb0c6749173a82c2b655debe5326141ae40d6bc70a WatchSource:0}: Error finding container d785ce62e296d0520fedbfbb0c6749173a82c2b655debe5326141ae40d6bc70a: Status 404 returned error can't find the container with id d785ce62e296d0520fedbfbb0c6749173a82c2b655debe5326141ae40d6bc70a Oct 08 07:01:58 crc kubenswrapper[4810]: I1008 07:01:58.367144 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6n45"] Oct 08 07:01:58 crc kubenswrapper[4810]: I1008 07:01:58.658055 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lt8wj" event={"ID":"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a","Type":"ContainerStarted","Data":"dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477"} Oct 08 07:01:58 crc kubenswrapper[4810]: I1008 07:01:58.660398 4810 generic.go:334] "Generic (PLEG): container finished" podID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerID="79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f" exitCode=0 Oct 08 07:01:58 crc kubenswrapper[4810]: I1008 07:01:58.660451 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6n45" event={"ID":"4694fb39-e621-47f0-9541-423a2f6f9e4b","Type":"ContainerDied","Data":"79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f"} Oct 08 07:01:58 crc kubenswrapper[4810]: I1008 07:01:58.660479 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6n45" event={"ID":"4694fb39-e621-47f0-9541-423a2f6f9e4b","Type":"ContainerStarted","Data":"d785ce62e296d0520fedbfbb0c6749173a82c2b655debe5326141ae40d6bc70a"} Oct 08 07:01:58 crc kubenswrapper[4810]: I1008 07:01:58.685501 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lt8wj" podStartSLOduration=2.204521307 podStartE2EDuration="3.685478121s" podCreationTimestamp="2025-10-08 07:01:55 +0000 UTC" firstStartedPulling="2025-10-08 07:01:56.642300368 +0000 UTC m=+1819.276740108" lastFinishedPulling="2025-10-08 07:01:58.123257182 +0000 UTC m=+1820.757696922" observedRunningTime="2025-10-08 07:01:58.68509997 +0000 UTC m=+1821.319539720" watchObservedRunningTime="2025-10-08 07:01:58.685478121 +0000 UTC m=+1821.319917871" Oct 08 07:02:00 crc kubenswrapper[4810]: I1008 07:02:00.677831 4810 generic.go:334] "Generic (PLEG): container finished" podID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerID="93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7" exitCode=0 Oct 08 07:02:00 crc kubenswrapper[4810]: I1008 07:02:00.677887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6n45" event={"ID":"4694fb39-e621-47f0-9541-423a2f6f9e4b","Type":"ContainerDied","Data":"93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7"} Oct 08 07:02:01 crc kubenswrapper[4810]: I1008 07:02:01.690087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6n45" event={"ID":"4694fb39-e621-47f0-9541-423a2f6f9e4b","Type":"ContainerStarted","Data":"70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9"} Oct 08 07:02:01 crc kubenswrapper[4810]: I1008 07:02:01.711048 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q6n45" podStartSLOduration=2.202314199 podStartE2EDuration="4.71102601s" podCreationTimestamp="2025-10-08 07:01:57 +0000 UTC" firstStartedPulling="2025-10-08 07:01:58.662399355 +0000 UTC m=+1821.296839095" lastFinishedPulling="2025-10-08 07:02:01.171111176 +0000 UTC m=+1823.805550906" observedRunningTime="2025-10-08 07:02:01.706886877 +0000 UTC m=+1824.341326627" watchObservedRunningTime="2025-10-08 07:02:01.71102601 +0000 UTC m=+1824.345465790" Oct 08 07:02:05 crc kubenswrapper[4810]: I1008 07:02:05.907576 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:02:05 crc kubenswrapper[4810]: I1008 07:02:05.910048 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:02:05 crc kubenswrapper[4810]: I1008 07:02:05.959639 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:02:06 crc kubenswrapper[4810]: I1008 07:02:06.805046 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:02:06 crc kubenswrapper[4810]: I1008 07:02:06.865288 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lt8wj"] Oct 08 07:02:08 crc kubenswrapper[4810]: I1008 07:02:08.102849 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:02:08 crc kubenswrapper[4810]: I1008 07:02:08.102903 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:02:08 crc kubenswrapper[4810]: I1008 07:02:08.171380 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:02:08 crc kubenswrapper[4810]: I1008 07:02:08.750103 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lt8wj" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="registry-server" containerID="cri-o://dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477" gracePeriod=2 Oct 08 07:02:08 crc kubenswrapper[4810]: I1008 07:02:08.841132 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.178814 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.298272 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-catalog-content\") pod \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.298441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vvh8\" (UniqueName: \"kubernetes.io/projected/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-kube-api-access-5vvh8\") pod \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.298515 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-utilities\") pod \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\" (UID: \"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a\") " Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.300117 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-utilities" (OuterVolumeSpecName: "utilities") pod "5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" (UID: "5cb66151-ce5c-4c0c-91dd-3da6167ebf2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.307498 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-kube-api-access-5vvh8" (OuterVolumeSpecName: "kube-api-access-5vvh8") pod "5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" (UID: "5cb66151-ce5c-4c0c-91dd-3da6167ebf2a"). InnerVolumeSpecName "kube-api-access-5vvh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.343538 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" (UID: "5cb66151-ce5c-4c0c-91dd-3da6167ebf2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.400552 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vvh8\" (UniqueName: \"kubernetes.io/projected/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-kube-api-access-5vvh8\") on node \"crc\" DevicePath \"\"" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.400661 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.400671 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.762061 4810 generic.go:334] "Generic (PLEG): container finished" podID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerID="dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477" exitCode=0 Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.762135 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lt8wj" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.762201 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lt8wj" event={"ID":"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a","Type":"ContainerDied","Data":"dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477"} Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.762276 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lt8wj" event={"ID":"5cb66151-ce5c-4c0c-91dd-3da6167ebf2a","Type":"ContainerDied","Data":"e504ae8dd91e9ffd2e41674d2fbff9a81913dba2e6b55ce0dfcbccafb6712650"} Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.762310 4810 scope.go:117] "RemoveContainer" containerID="dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.794385 4810 scope.go:117] "RemoveContainer" containerID="bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.811286 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lt8wj"] Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.817786 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lt8wj"] Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.836707 4810 scope.go:117] "RemoveContainer" containerID="65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.871673 4810 scope.go:117] "RemoveContainer" containerID="dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477" Oct 08 07:02:09 crc kubenswrapper[4810]: E1008 07:02:09.873226 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477\": container with ID starting with dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477 not found: ID does not exist" containerID="dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.873291 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477"} err="failed to get container status \"dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477\": rpc error: code = NotFound desc = could not find container \"dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477\": container with ID starting with dcbf5edd7874b31493b57a4f50da81c5e162904019c5faf007e97c69154d5477 not found: ID does not exist" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.873327 4810 scope.go:117] "RemoveContainer" containerID="bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb" Oct 08 07:02:09 crc kubenswrapper[4810]: E1008 07:02:09.873915 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb\": container with ID starting with bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb not found: ID does not exist" containerID="bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.873956 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb"} err="failed to get container status \"bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb\": rpc error: code = NotFound desc = could not find container \"bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb\": container with ID starting with bdfe5fcda8486bb91a217f0f47fd415f25da21c8a9efdcf8349f833a9558f1eb not found: ID does not exist" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.873998 4810 scope.go:117] "RemoveContainer" containerID="65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e" Oct 08 07:02:09 crc kubenswrapper[4810]: E1008 07:02:09.874412 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e\": container with ID starting with 65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e not found: ID does not exist" containerID="65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e" Oct 08 07:02:09 crc kubenswrapper[4810]: I1008 07:02:09.874452 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e"} err="failed to get container status \"65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e\": rpc error: code = NotFound desc = could not find container \"65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e\": container with ID starting with 65d546cc5022498beb834d940bfa262583b95b207b7bdafd9750576f6108c63e not found: ID does not exist" Oct 08 07:02:10 crc kubenswrapper[4810]: I1008 07:02:10.006560 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6n45"] Oct 08 07:02:10 crc kubenswrapper[4810]: I1008 07:02:10.087506 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" path="/var/lib/kubelet/pods/5cb66151-ce5c-4c0c-91dd-3da6167ebf2a/volumes" Oct 08 07:02:10 crc kubenswrapper[4810]: I1008 07:02:10.772873 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q6n45" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="registry-server" containerID="cri-o://70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9" gracePeriod=2 Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.197983 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.226682 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6vlt\" (UniqueName: \"kubernetes.io/projected/4694fb39-e621-47f0-9541-423a2f6f9e4b-kube-api-access-n6vlt\") pod \"4694fb39-e621-47f0-9541-423a2f6f9e4b\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.226743 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-utilities\") pod \"4694fb39-e621-47f0-9541-423a2f6f9e4b\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.226782 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-catalog-content\") pod \"4694fb39-e621-47f0-9541-423a2f6f9e4b\" (UID: \"4694fb39-e621-47f0-9541-423a2f6f9e4b\") " Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.227584 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-utilities" (OuterVolumeSpecName: "utilities") pod "4694fb39-e621-47f0-9541-423a2f6f9e4b" (UID: "4694fb39-e621-47f0-9541-423a2f6f9e4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.239606 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4694fb39-e621-47f0-9541-423a2f6f9e4b-kube-api-access-n6vlt" (OuterVolumeSpecName: "kube-api-access-n6vlt") pod "4694fb39-e621-47f0-9541-423a2f6f9e4b" (UID: "4694fb39-e621-47f0-9541-423a2f6f9e4b"). InnerVolumeSpecName "kube-api-access-n6vlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.240372 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4694fb39-e621-47f0-9541-423a2f6f9e4b" (UID: "4694fb39-e621-47f0-9541-423a2f6f9e4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.328212 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6vlt\" (UniqueName: \"kubernetes.io/projected/4694fb39-e621-47f0-9541-423a2f6f9e4b-kube-api-access-n6vlt\") on node \"crc\" DevicePath \"\"" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.328246 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.328255 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4694fb39-e621-47f0-9541-423a2f6f9e4b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.781313 4810 generic.go:334] "Generic (PLEG): container finished" podID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerID="70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9" exitCode=0 Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.781351 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q6n45" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.781368 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6n45" event={"ID":"4694fb39-e621-47f0-9541-423a2f6f9e4b","Type":"ContainerDied","Data":"70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9"} Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.781438 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q6n45" event={"ID":"4694fb39-e621-47f0-9541-423a2f6f9e4b","Type":"ContainerDied","Data":"d785ce62e296d0520fedbfbb0c6749173a82c2b655debe5326141ae40d6bc70a"} Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.781469 4810 scope.go:117] "RemoveContainer" containerID="70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.801891 4810 scope.go:117] "RemoveContainer" containerID="93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.815298 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6n45"] Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.824607 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q6n45"] Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.827700 4810 scope.go:117] "RemoveContainer" containerID="79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.848760 4810 scope.go:117] "RemoveContainer" containerID="70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9" Oct 08 07:02:11 crc kubenswrapper[4810]: E1008 07:02:11.849252 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9\": container with ID starting with 70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9 not found: ID does not exist" containerID="70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.849314 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9"} err="failed to get container status \"70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9\": rpc error: code = NotFound desc = could not find container \"70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9\": container with ID starting with 70b654f7edce1720ff9e8daaaacde1303f4adcebd2c09c709c4ddb972d99e2b9 not found: ID does not exist" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.849355 4810 scope.go:117] "RemoveContainer" containerID="93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7" Oct 08 07:02:11 crc kubenswrapper[4810]: E1008 07:02:11.849825 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7\": container with ID starting with 93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7 not found: ID does not exist" containerID="93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.849866 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7"} err="failed to get container status \"93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7\": rpc error: code = NotFound desc = could not find container \"93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7\": container with ID starting with 93a8068467da5b29bbafca05e188c2f54f3c4eee06373f6ba44147ae307b4ad7 not found: ID does not exist" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.849896 4810 scope.go:117] "RemoveContainer" containerID="79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f" Oct 08 07:02:11 crc kubenswrapper[4810]: E1008 07:02:11.850141 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f\": container with ID starting with 79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f not found: ID does not exist" containerID="79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f" Oct 08 07:02:11 crc kubenswrapper[4810]: I1008 07:02:11.850183 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f"} err="failed to get container status \"79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f\": rpc error: code = NotFound desc = could not find container \"79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f\": container with ID starting with 79eaa258b18f72d6804646331803b107628345db0261e62634c9cad82bec9c4f not found: ID does not exist" Oct 08 07:02:12 crc kubenswrapper[4810]: I1008 07:02:12.087001 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" path="/var/lib/kubelet/pods/4694fb39-e621-47f0-9541-423a2f6f9e4b/volumes" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.717329 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kzxbp/must-gather-2zfqm"] Oct 08 07:03:05 crc kubenswrapper[4810]: E1008 07:03:05.718662 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="extract-content" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.718687 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="extract-content" Oct 08 07:03:05 crc kubenswrapper[4810]: E1008 07:03:05.718772 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="extract-utilities" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.718790 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="extract-utilities" Oct 08 07:03:05 crc kubenswrapper[4810]: E1008 07:03:05.718817 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="registry-server" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.718829 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="registry-server" Oct 08 07:03:05 crc kubenswrapper[4810]: E1008 07:03:05.718845 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="extract-utilities" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.718855 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="extract-utilities" Oct 08 07:03:05 crc kubenswrapper[4810]: E1008 07:03:05.718869 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="registry-server" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.718880 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="registry-server" Oct 08 07:03:05 crc kubenswrapper[4810]: E1008 07:03:05.718902 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="extract-content" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.718915 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="extract-content" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.719229 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb66151-ce5c-4c0c-91dd-3da6167ebf2a" containerName="registry-server" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.719265 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4694fb39-e621-47f0-9541-423a2f6f9e4b" containerName="registry-server" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.720517 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.724007 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kzxbp"/"kube-root-ca.crt" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.724435 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kzxbp"/"openshift-service-ca.crt" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.730585 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kzxbp/must-gather-2zfqm"] Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.780462 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7168350-094e-408e-9e5f-472928c39636-must-gather-output\") pod \"must-gather-2zfqm\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.780506 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffq2j\" (UniqueName: \"kubernetes.io/projected/c7168350-094e-408e-9e5f-472928c39636-kube-api-access-ffq2j\") pod \"must-gather-2zfqm\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.881367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7168350-094e-408e-9e5f-472928c39636-must-gather-output\") pod \"must-gather-2zfqm\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.881410 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffq2j\" (UniqueName: \"kubernetes.io/projected/c7168350-094e-408e-9e5f-472928c39636-kube-api-access-ffq2j\") pod \"must-gather-2zfqm\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.882078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7168350-094e-408e-9e5f-472928c39636-must-gather-output\") pod \"must-gather-2zfqm\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:05 crc kubenswrapper[4810]: I1008 07:03:05.909091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffq2j\" (UniqueName: \"kubernetes.io/projected/c7168350-094e-408e-9e5f-472928c39636-kube-api-access-ffq2j\") pod \"must-gather-2zfqm\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:06 crc kubenswrapper[4810]: I1008 07:03:06.046781 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:03:06 crc kubenswrapper[4810]: I1008 07:03:06.521272 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kzxbp/must-gather-2zfqm"] Oct 08 07:03:07 crc kubenswrapper[4810]: I1008 07:03:07.220476 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" event={"ID":"c7168350-094e-408e-9e5f-472928c39636","Type":"ContainerStarted","Data":"dbdb0bba45d75cf28e832f6b8e1073486b8e9c33f59ec7087345bd11816f9cbf"} Oct 08 07:03:11 crc kubenswrapper[4810]: I1008 07:03:11.247673 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" event={"ID":"c7168350-094e-408e-9e5f-472928c39636","Type":"ContainerStarted","Data":"9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838"} Oct 08 07:03:11 crc kubenswrapper[4810]: I1008 07:03:11.248237 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" event={"ID":"c7168350-094e-408e-9e5f-472928c39636","Type":"ContainerStarted","Data":"96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141"} Oct 08 07:03:11 crc kubenswrapper[4810]: I1008 07:03:11.265387 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" podStartSLOduration=2.034401361 podStartE2EDuration="6.265364078s" podCreationTimestamp="2025-10-08 07:03:05 +0000 UTC" firstStartedPulling="2025-10-08 07:03:06.527479214 +0000 UTC m=+1889.161918954" lastFinishedPulling="2025-10-08 07:03:10.758441901 +0000 UTC m=+1893.392881671" observedRunningTime="2025-10-08 07:03:11.259304491 +0000 UTC m=+1893.893744241" watchObservedRunningTime="2025-10-08 07:03:11.265364078 +0000 UTC m=+1893.899803818" Oct 08 07:04:02 crc kubenswrapper[4810]: I1008 07:04:02.530730 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb_bd9febeb-3b39-4d90-9c95-76f86ac8b87e/util/0.log" Oct 08 07:04:02 crc kubenswrapper[4810]: I1008 07:04:02.677737 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb_bd9febeb-3b39-4d90-9c95-76f86ac8b87e/util/0.log" Oct 08 07:04:02 crc kubenswrapper[4810]: I1008 07:04:02.707463 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb_bd9febeb-3b39-4d90-9c95-76f86ac8b87e/pull/0.log" Oct 08 07:04:02 crc kubenswrapper[4810]: I1008 07:04:02.787455 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb_bd9febeb-3b39-4d90-9c95-76f86ac8b87e/pull/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.002753 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb_bd9febeb-3b39-4d90-9c95-76f86ac8b87e/util/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.019955 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb_bd9febeb-3b39-4d90-9c95-76f86ac8b87e/extract/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.022555 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_36122a92e6b12339df3c30325af8b6d046b31b95ba79e8e607770040954ltdb_bd9febeb-3b39-4d90-9c95-76f86ac8b87e/pull/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.228434 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f56ff694-tlndl_3fe805b1-a990-4683-8549-f73ebe9d850c/kube-rbac-proxy/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.262115 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f56ff694-tlndl_3fe805b1-a990-4683-8549-f73ebe9d850c/manager/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.522816 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-2lq4z_2c5ab097-4362-44a0-8f0a-a2a2483851f4/kube-rbac-proxy/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.695715 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-2lq4z_2c5ab097-4362-44a0-8f0a-a2a2483851f4/manager/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.713075 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-h4wk6_75590e1d-5795-4ebe-afd0-9abb818efd63/manager/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.730640 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-h4wk6_75590e1d-5795-4ebe-afd0-9abb818efd63/kube-rbac-proxy/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.879027 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-fd648f65-nwwg7_0bebe3c8-1035-40b4-b533-4c214423f6b2/kube-rbac-proxy/0.log" Oct 08 07:04:03 crc kubenswrapper[4810]: I1008 07:04:03.928755 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-fd648f65-nwwg7_0bebe3c8-1035-40b4-b533-4c214423f6b2/manager/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.121554 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7ccfc8cf49-ntp7k_6fc03514-36ec-4d41-8ad1-306ff2767a2f/kube-rbac-proxy/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.136152 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7ccfc8cf49-ntp7k_6fc03514-36ec-4d41-8ad1-306ff2767a2f/manager/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.190377 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b477879bc-8jhkm_72a051e1-5079-45b0-b3d8-0592f872c85a/kube-rbac-proxy/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.378384 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b477879bc-8jhkm_72a051e1-5079-45b0-b3d8-0592f872c85a/manager/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.492632 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-dqjfx_0b266b2d-a60b-4990-b720-800d655a4b5f/kube-rbac-proxy/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.643423 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-dqjfx_0b266b2d-a60b-4990-b720-800d655a4b5f/manager/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.685091 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5467f8988c-s4m9g_8181e145-9c88-47f0-a9a8-dc03947b75ef/manager/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.826748 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5467f8988c-s4m9g_8181e145-9c88-47f0-a9a8-dc03947b75ef/kube-rbac-proxy/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.886989 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b84cc7657-9gdl9_24b64e0e-8eb4-4235-9086-8f1f61f2bd8e/kube-rbac-proxy/0.log" Oct 08 07:04:04 crc kubenswrapper[4810]: I1008 07:04:04.928909 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b84cc7657-9gdl9_24b64e0e-8eb4-4235-9086-8f1f61f2bd8e/manager/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.078419 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-657k2_e8deb064-ce04-4e12-ab87-e4853e499c20/kube-rbac-proxy/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.088605 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-657k2_e8deb064-ce04-4e12-ab87-e4853e499c20/manager/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.268157 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-ckvcg_c19839d0-64b9-4f39-a4e2-40cdc62bd9b1/kube-rbac-proxy/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.298257 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-ckvcg_c19839d0-64b9-4f39-a4e2-40cdc62bd9b1/manager/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.370641 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-r9g7m_09c37aa7-fb5d-4984-9f32-86a4a323ea49/kube-rbac-proxy/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.475352 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-r9g7m_09c37aa7-fb5d-4984-9f32-86a4a323ea49/manager/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.536812 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-8cmwf_681220bd-766a-4d1e-970b-662b4c64aaf6/kube-rbac-proxy/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.646814 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-8cmwf_681220bd-766a-4d1e-970b-662b4c64aaf6/manager/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.730576 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-rtmss_413157c3-1aa7-425d-98ae-e44a7c19b425/kube-rbac-proxy/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.735733 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-rtmss_413157c3-1aa7-425d-98ae-e44a7c19b425/manager/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.874609 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6875c66686lr5fk_f164aece-1b13-46b7-9da2-16967eb942f5/kube-rbac-proxy/0.log" Oct 08 07:04:05 crc kubenswrapper[4810]: I1008 07:04:05.931355 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6875c66686lr5fk_f164aece-1b13-46b7-9da2-16967eb942f5/manager/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.053477 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6bfd56c677-ljndg_ea4ec203-527d-4deb-94c5-068053bbb7ee/kube-rbac-proxy/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.124754 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b6d857f89-wtlfl_71fb758c-689c-4098-b73a-e45b4bf6db60/kube-rbac-proxy/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.323527 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-b6d857f89-wtlfl_71fb758c-689c-4098-b73a-e45b4bf6db60/operator/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.364324 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-q9z86_402bba51-e1a6-4f79-823f-864c2d99a0e6/registry-server/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.582604 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54d485fd9-29jxc_cb1fd16d-a191-4086-9a18-37b74183042d/kube-rbac-proxy/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.590729 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54d485fd9-29jxc_cb1fd16d-a191-4086-9a18-37b74183042d/manager/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.641695 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-mnzm7_568c0ce9-12f3-4b04-9131-83771329a83e/kube-rbac-proxy/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.776715 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6bfd56c677-ljndg_ea4ec203-527d-4deb-94c5-068053bbb7ee/manager/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.788516 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-mnzm7_568c0ce9-12f3-4b04-9131-83771329a83e/manager/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.893742 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-xxkpw_133a94f9-3f74-42f7-95cf-dd0b193b6626/operator/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.981434 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-qfp99_914ce43c-e528-4921-b1b4-aa47cdf4b004/manager/0.log" Oct 08 07:04:06 crc kubenswrapper[4810]: I1008 07:04:06.989751 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-qfp99_914ce43c-e528-4921-b1b4-aa47cdf4b004/kube-rbac-proxy/0.log" Oct 08 07:04:07 crc kubenswrapper[4810]: I1008 07:04:07.070907 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-274pz_826e673f-b91d-4604-959f-098bcc2e6dcd/kube-rbac-proxy/0.log" Oct 08 07:04:07 crc kubenswrapper[4810]: I1008 07:04:07.140448 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-274pz_826e673f-b91d-4604-959f-098bcc2e6dcd/manager/0.log" Oct 08 07:04:07 crc kubenswrapper[4810]: I1008 07:04:07.222956 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-ldhbx_17a2f9a4-8153-490e-be01-e452b10638e9/kube-rbac-proxy/0.log" Oct 08 07:04:07 crc kubenswrapper[4810]: I1008 07:04:07.304089 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-ldhbx_17a2f9a4-8153-490e-be01-e452b10638e9/manager/0.log" Oct 08 07:04:07 crc kubenswrapper[4810]: I1008 07:04:07.351506 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-ssc8r_069894d9-bd50-4209-9df7-d941d44f0341/kube-rbac-proxy/0.log" Oct 08 07:04:07 crc kubenswrapper[4810]: I1008 07:04:07.356528 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-ssc8r_069894d9-bd50-4209-9df7-d941d44f0341/manager/0.log" Oct 08 07:04:19 crc kubenswrapper[4810]: I1008 07:04:19.431790 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 07:04:19 crc kubenswrapper[4810]: I1008 07:04:19.432809 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 07:04:21 crc kubenswrapper[4810]: I1008 07:04:21.735490 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-q75tf_12e1e52f-d7f4-4568-a0ac-79ce8102ff96/control-plane-machine-set-operator/0.log" Oct 08 07:04:21 crc kubenswrapper[4810]: I1008 07:04:21.921502 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jn2mz_80088141-deec-4bdb-b54a-10fce37e97a0/machine-api-operator/0.log" Oct 08 07:04:21 crc kubenswrapper[4810]: I1008 07:04:21.944038 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jn2mz_80088141-deec-4bdb-b54a-10fce37e97a0/kube-rbac-proxy/0.log" Oct 08 07:04:32 crc kubenswrapper[4810]: I1008 07:04:32.558745 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-ptlfz_11ae85d1-fe3e-49ec-a984-1dc8e35050d6/cert-manager-controller/0.log" Oct 08 07:04:32 crc kubenswrapper[4810]: I1008 07:04:32.698022 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-cfv7r_82648b77-ff5f-4c3b-93b9-3a860b9f4149/cert-manager-cainjector/0.log" Oct 08 07:04:32 crc kubenswrapper[4810]: I1008 07:04:32.756155 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-7chth_0d166b56-ad1e-44dc-96b3-58d1d3116642/cert-manager-webhook/0.log" Oct 08 07:04:43 crc kubenswrapper[4810]: I1008 07:04:43.856934 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-62h9n_b2cac821-64bd-4d46-8d3e-e8a51fd9d27f/nmstate-console-plugin/0.log" Oct 08 07:04:44 crc kubenswrapper[4810]: I1008 07:04:44.018889 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rqsl2_bfcded37-ea7d-4d11-abc2-d66055730182/nmstate-handler/0.log" Oct 08 07:04:44 crc kubenswrapper[4810]: I1008 07:04:44.138480 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-z68nd_c1e3d23b-7985-42b6-866b-fd5a93c9921f/kube-rbac-proxy/0.log" Oct 08 07:04:44 crc kubenswrapper[4810]: I1008 07:04:44.147083 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-z68nd_c1e3d23b-7985-42b6-866b-fd5a93c9921f/nmstate-metrics/0.log" Oct 08 07:04:44 crc kubenswrapper[4810]: I1008 07:04:44.250392 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-c75hl_65727c84-7a72-4ffb-91bc-93e2242d360a/nmstate-operator/0.log" Oct 08 07:04:44 crc kubenswrapper[4810]: I1008 07:04:44.341266 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-8x6qm_9d3fe3d3-473d-457f-9e5a-d71de5d1e4ba/nmstate-webhook/0.log" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.421297 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p5mj6"] Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.423992 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.436552 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5mj6"] Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.531958 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-utilities\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.532120 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-catalog-content\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.532156 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njgl4\" (UniqueName: \"kubernetes.io/projected/d9330d29-ee52-4255-bb66-2871eec3e612-kube-api-access-njgl4\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.633805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-utilities\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.633924 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-catalog-content\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.633953 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njgl4\" (UniqueName: \"kubernetes.io/projected/d9330d29-ee52-4255-bb66-2871eec3e612-kube-api-access-njgl4\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.634548 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-utilities\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.634600 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-catalog-content\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.656162 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njgl4\" (UniqueName: \"kubernetes.io/projected/d9330d29-ee52-4255-bb66-2871eec3e612-kube-api-access-njgl4\") pod \"redhat-operators-p5mj6\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:46 crc kubenswrapper[4810]: I1008 07:04:46.749883 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:47 crc kubenswrapper[4810]: I1008 07:04:47.016897 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5mj6"] Oct 08 07:04:47 crc kubenswrapper[4810]: I1008 07:04:47.933879 4810 generic.go:334] "Generic (PLEG): container finished" podID="d9330d29-ee52-4255-bb66-2871eec3e612" containerID="268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae" exitCode=0 Oct 08 07:04:47 crc kubenswrapper[4810]: I1008 07:04:47.933922 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5mj6" event={"ID":"d9330d29-ee52-4255-bb66-2871eec3e612","Type":"ContainerDied","Data":"268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae"} Oct 08 07:04:47 crc kubenswrapper[4810]: I1008 07:04:47.934221 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5mj6" event={"ID":"d9330d29-ee52-4255-bb66-2871eec3e612","Type":"ContainerStarted","Data":"fce504e85e82561bd925f4025efe4fafb73e3d2e79777040eb8db83c644230e7"} Oct 08 07:04:49 crc kubenswrapper[4810]: I1008 07:04:49.431528 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 07:04:49 crc kubenswrapper[4810]: I1008 07:04:49.431850 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 07:04:49 crc kubenswrapper[4810]: I1008 07:04:49.947176 4810 generic.go:334] "Generic (PLEG): container finished" podID="d9330d29-ee52-4255-bb66-2871eec3e612" containerID="8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c" exitCode=0 Oct 08 07:04:49 crc kubenswrapper[4810]: I1008 07:04:49.947604 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5mj6" event={"ID":"d9330d29-ee52-4255-bb66-2871eec3e612","Type":"ContainerDied","Data":"8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c"} Oct 08 07:04:51 crc kubenswrapper[4810]: I1008 07:04:51.963112 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5mj6" event={"ID":"d9330d29-ee52-4255-bb66-2871eec3e612","Type":"ContainerStarted","Data":"2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665"} Oct 08 07:04:51 crc kubenswrapper[4810]: I1008 07:04:51.981820 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p5mj6" podStartSLOduration=3.089494864 podStartE2EDuration="5.981803674s" podCreationTimestamp="2025-10-08 07:04:46 +0000 UTC" firstStartedPulling="2025-10-08 07:04:47.936074207 +0000 UTC m=+1990.570513947" lastFinishedPulling="2025-10-08 07:04:50.828382997 +0000 UTC m=+1993.462822757" observedRunningTime="2025-10-08 07:04:51.981609339 +0000 UTC m=+1994.616049069" watchObservedRunningTime="2025-10-08 07:04:51.981803674 +0000 UTC m=+1994.616243404" Oct 08 07:04:56 crc kubenswrapper[4810]: I1008 07:04:56.750169 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:56 crc kubenswrapper[4810]: I1008 07:04:56.750578 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:56 crc kubenswrapper[4810]: I1008 07:04:56.797039 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.034366 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.084641 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5mj6"] Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.456241 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-4xvt4_f95f0a34-32e5-4bd3-a2f7-73afabbfc182/kube-rbac-proxy/0.log" Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.740520 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-frr-files/0.log" Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.910636 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-frr-files/0.log" Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.917374 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-4xvt4_f95f0a34-32e5-4bd3-a2f7-73afabbfc182/controller/0.log" Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.959051 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-metrics/0.log" Oct 08 07:04:57 crc kubenswrapper[4810]: I1008 07:04:57.979047 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-reloader/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.097542 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-reloader/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.227902 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-reloader/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.235466 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-frr-files/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.248586 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-metrics/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.318518 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-metrics/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.430035 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-metrics/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.454570 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-reloader/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.466185 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/cp-frr-files/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.519648 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/controller/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.621605 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/frr-metrics/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.692984 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/kube-rbac-proxy/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.713793 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/kube-rbac-proxy-frr/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.831801 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/reloader/0.log" Oct 08 07:04:58 crc kubenswrapper[4810]: I1008 07:04:58.955774 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-qh2dm_41795b92-c731-421a-bf39-f3d48e99b84c/frr-k8s-webhook-server/0.log" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.005692 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p5mj6" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="registry-server" containerID="cri-o://2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665" gracePeriod=2 Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.146287 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-85f4bbfddc-bx2db_1cf28dd8-f466-44e0-b814-b956dda73d9c/manager/0.log" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.242299 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-75dc598b55-b4qjh_08e95f35-af58-4c5a-9d5c-3fb6dd406a55/webhook-server/0.log" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.412138 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.456902 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ffrq4_30febcf0-5a02-4efd-b2c7-f1fadc2f0f51/kube-rbac-proxy/0.log" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.513329 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-utilities\") pod \"d9330d29-ee52-4255-bb66-2871eec3e612\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.513444 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njgl4\" (UniqueName: \"kubernetes.io/projected/d9330d29-ee52-4255-bb66-2871eec3e612-kube-api-access-njgl4\") pod \"d9330d29-ee52-4255-bb66-2871eec3e612\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.513478 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-catalog-content\") pod \"d9330d29-ee52-4255-bb66-2871eec3e612\" (UID: \"d9330d29-ee52-4255-bb66-2871eec3e612\") " Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.522753 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-utilities" (OuterVolumeSpecName: "utilities") pod "d9330d29-ee52-4255-bb66-2871eec3e612" (UID: "d9330d29-ee52-4255-bb66-2871eec3e612"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.532155 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9330d29-ee52-4255-bb66-2871eec3e612-kube-api-access-njgl4" (OuterVolumeSpecName: "kube-api-access-njgl4") pod "d9330d29-ee52-4255-bb66-2871eec3e612" (UID: "d9330d29-ee52-4255-bb66-2871eec3e612"). InnerVolumeSpecName "kube-api-access-njgl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.615587 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.615617 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njgl4\" (UniqueName: \"kubernetes.io/projected/d9330d29-ee52-4255-bb66-2871eec3e612-kube-api-access-njgl4\") on node \"crc\" DevicePath \"\"" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.636074 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9330d29-ee52-4255-bb66-2871eec3e612" (UID: "d9330d29-ee52-4255-bb66-2871eec3e612"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.717008 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9330d29-ee52-4255-bb66-2871eec3e612-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.924300 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ffrq4_30febcf0-5a02-4efd-b2c7-f1fadc2f0f51/speaker/0.log" Oct 08 07:04:59 crc kubenswrapper[4810]: I1008 07:04:59.964883 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ncsk2_6c32cd68-ec2a-4134-a7f4-da9466733bc1/frr/0.log" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.014949 4810 generic.go:334] "Generic (PLEG): container finished" podID="d9330d29-ee52-4255-bb66-2871eec3e612" containerID="2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665" exitCode=0 Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.015007 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5mj6" event={"ID":"d9330d29-ee52-4255-bb66-2871eec3e612","Type":"ContainerDied","Data":"2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665"} Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.015036 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5mj6" event={"ID":"d9330d29-ee52-4255-bb66-2871eec3e612","Type":"ContainerDied","Data":"fce504e85e82561bd925f4025efe4fafb73e3d2e79777040eb8db83c644230e7"} Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.015055 4810 scope.go:117] "RemoveContainer" containerID="2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.015082 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5mj6" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.036340 4810 scope.go:117] "RemoveContainer" containerID="8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.053080 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5mj6"] Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.066294 4810 scope.go:117] "RemoveContainer" containerID="268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.103308 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p5mj6"] Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.104302 4810 scope.go:117] "RemoveContainer" containerID="2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665" Oct 08 07:05:00 crc kubenswrapper[4810]: E1008 07:05:00.104708 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665\": container with ID starting with 2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665 not found: ID does not exist" containerID="2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.104740 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665"} err="failed to get container status \"2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665\": rpc error: code = NotFound desc = could not find container \"2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665\": container with ID starting with 2308b5dfe74bf8dabbc6d8c2beecb58adf8f57bea948cf3bc00d1bac086fc665 not found: ID does not exist" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.104761 4810 scope.go:117] "RemoveContainer" containerID="8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c" Oct 08 07:05:00 crc kubenswrapper[4810]: E1008 07:05:00.105394 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c\": container with ID starting with 8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c not found: ID does not exist" containerID="8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.105433 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c"} err="failed to get container status \"8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c\": rpc error: code = NotFound desc = could not find container \"8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c\": container with ID starting with 8bf6879ae288965edac348e3ef52e6577b4e51cdceb2049338659afc3bde150c not found: ID does not exist" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.105447 4810 scope.go:117] "RemoveContainer" containerID="268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae" Oct 08 07:05:00 crc kubenswrapper[4810]: E1008 07:05:00.106491 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae\": container with ID starting with 268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae not found: ID does not exist" containerID="268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae" Oct 08 07:05:00 crc kubenswrapper[4810]: I1008 07:05:00.106514 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae"} err="failed to get container status \"268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae\": rpc error: code = NotFound desc = could not find container \"268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae\": container with ID starting with 268f57abd36f95456eb0646e236ba8075fb6a5f30a196871cfc0cf19b64d3dae not found: ID does not exist" Oct 08 07:05:02 crc kubenswrapper[4810]: I1008 07:05:02.082378 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" path="/var/lib/kubelet/pods/d9330d29-ee52-4255-bb66-2871eec3e612/volumes" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.182560 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z_497ef608-88af-4370-8886-c5377ec5b1e8/util/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.346011 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z_497ef608-88af-4370-8886-c5377ec5b1e8/util/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.393319 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z_497ef608-88af-4370-8886-c5377ec5b1e8/pull/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.394635 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z_497ef608-88af-4370-8886-c5377ec5b1e8/pull/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.576731 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z_497ef608-88af-4370-8886-c5377ec5b1e8/util/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.614943 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z_497ef608-88af-4370-8886-c5377ec5b1e8/pull/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.624947 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zrt7z_497ef608-88af-4370-8886-c5377ec5b1e8/extract/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.733483 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk_a378b60b-7299-43c4-b159-5b5c54e41b11/util/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.894768 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk_a378b60b-7299-43c4-b159-5b5c54e41b11/util/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.930837 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk_a378b60b-7299-43c4-b159-5b5c54e41b11/pull/0.log" Oct 08 07:05:11 crc kubenswrapper[4810]: I1008 07:05:11.959480 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk_a378b60b-7299-43c4-b159-5b5c54e41b11/pull/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.082248 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk_a378b60b-7299-43c4-b159-5b5c54e41b11/util/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.083310 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk_a378b60b-7299-43c4-b159-5b5c54e41b11/pull/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.084198 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2s6wrk_a378b60b-7299-43c4-b159-5b5c54e41b11/extract/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.249733 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cv58h_90532ac2-5faa-4f71-b4ea-4c9af34d0199/extract-utilities/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.422428 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cv58h_90532ac2-5faa-4f71-b4ea-4c9af34d0199/extract-content/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.427172 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cv58h_90532ac2-5faa-4f71-b4ea-4c9af34d0199/extract-content/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.438178 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cv58h_90532ac2-5faa-4f71-b4ea-4c9af34d0199/extract-utilities/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.567503 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cv58h_90532ac2-5faa-4f71-b4ea-4c9af34d0199/extract-utilities/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.575306 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cv58h_90532ac2-5faa-4f71-b4ea-4c9af34d0199/extract-content/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.736102 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl5kz_2798b3c2-26d6-41e1-a3a4-ab10b5af84be/extract-utilities/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.900414 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cv58h_90532ac2-5faa-4f71-b4ea-4c9af34d0199/registry-server/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.978908 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl5kz_2798b3c2-26d6-41e1-a3a4-ab10b5af84be/extract-content/0.log" Oct 08 07:05:12 crc kubenswrapper[4810]: I1008 07:05:12.987982 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl5kz_2798b3c2-26d6-41e1-a3a4-ab10b5af84be/extract-utilities/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.016196 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl5kz_2798b3c2-26d6-41e1-a3a4-ab10b5af84be/extract-content/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.165052 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl5kz_2798b3c2-26d6-41e1-a3a4-ab10b5af84be/extract-content/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.192865 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl5kz_2798b3c2-26d6-41e1-a3a4-ab10b5af84be/extract-utilities/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.350229 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w_93a3ddd7-2bd3-4b14-8334-faddbd4cd42c/util/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.469481 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dl5kz_2798b3c2-26d6-41e1-a3a4-ab10b5af84be/registry-server/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.658632 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w_93a3ddd7-2bd3-4b14-8334-faddbd4cd42c/pull/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.659374 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w_93a3ddd7-2bd3-4b14-8334-faddbd4cd42c/pull/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.727999 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w_93a3ddd7-2bd3-4b14-8334-faddbd4cd42c/util/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.802164 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w_93a3ddd7-2bd3-4b14-8334-faddbd4cd42c/util/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.826438 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w_93a3ddd7-2bd3-4b14-8334-faddbd4cd42c/pull/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.864771 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c8xv9w_93a3ddd7-2bd3-4b14-8334-faddbd4cd42c/extract/0.log" Oct 08 07:05:13 crc kubenswrapper[4810]: I1008 07:05:13.981126 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-k4b69_2f9ae171-32b7-4ca1-8d91-399067439899/marketplace-operator/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.068453 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zv7br_f658954e-7efb-4e24-95fa-b5191316ef0e/extract-utilities/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.287275 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zv7br_f658954e-7efb-4e24-95fa-b5191316ef0e/extract-utilities/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.291780 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zv7br_f658954e-7efb-4e24-95fa-b5191316ef0e/extract-content/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.323834 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zv7br_f658954e-7efb-4e24-95fa-b5191316ef0e/extract-content/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.504358 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zv7br_f658954e-7efb-4e24-95fa-b5191316ef0e/extract-content/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.572778 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zv7br_f658954e-7efb-4e24-95fa-b5191316ef0e/extract-utilities/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.661320 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zv7br_f658954e-7efb-4e24-95fa-b5191316ef0e/registry-server/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.712698 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jxmtt_28a0d2ff-103e-48be-a7e7-db75c60980f1/extract-utilities/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.942020 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jxmtt_28a0d2ff-103e-48be-a7e7-db75c60980f1/extract-content/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.955170 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jxmtt_28a0d2ff-103e-48be-a7e7-db75c60980f1/extract-content/0.log" Oct 08 07:05:14 crc kubenswrapper[4810]: I1008 07:05:14.968358 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jxmtt_28a0d2ff-103e-48be-a7e7-db75c60980f1/extract-utilities/0.log" Oct 08 07:05:15 crc kubenswrapper[4810]: I1008 07:05:15.149911 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jxmtt_28a0d2ff-103e-48be-a7e7-db75c60980f1/extract-utilities/0.log" Oct 08 07:05:15 crc kubenswrapper[4810]: I1008 07:05:15.165418 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jxmtt_28a0d2ff-103e-48be-a7e7-db75c60980f1/extract-content/0.log" Oct 08 07:05:15 crc kubenswrapper[4810]: I1008 07:05:15.456062 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jxmtt_28a0d2ff-103e-48be-a7e7-db75c60980f1/registry-server/0.log" Oct 08 07:05:19 crc kubenswrapper[4810]: I1008 07:05:19.431776 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 07:05:19 crc kubenswrapper[4810]: I1008 07:05:19.432265 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 07:05:19 crc kubenswrapper[4810]: I1008 07:05:19.432340 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" Oct 08 07:05:19 crc kubenswrapper[4810]: I1008 07:05:19.433258 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d02be0e8f536b325c81ec0c82d72faaf3765230f22216c32727e84a1825af9fe"} pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 07:05:19 crc kubenswrapper[4810]: I1008 07:05:19.433541 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" containerID="cri-o://d02be0e8f536b325c81ec0c82d72faaf3765230f22216c32727e84a1825af9fe" gracePeriod=600 Oct 08 07:05:20 crc kubenswrapper[4810]: I1008 07:05:20.178420 4810 generic.go:334] "Generic (PLEG): container finished" podID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerID="d02be0e8f536b325c81ec0c82d72faaf3765230f22216c32727e84a1825af9fe" exitCode=0 Oct 08 07:05:20 crc kubenswrapper[4810]: I1008 07:05:20.178473 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerDied","Data":"d02be0e8f536b325c81ec0c82d72faaf3765230f22216c32727e84a1825af9fe"} Oct 08 07:05:20 crc kubenswrapper[4810]: I1008 07:05:20.179030 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" event={"ID":"2b5ec8eb-9a64-46ae-b881-535323e3b686","Type":"ContainerStarted","Data":"308729fd12b5669c3eff02e71d7b8234d59fa1f9e1ec05fffd7636732663381f"} Oct 08 07:05:20 crc kubenswrapper[4810]: I1008 07:05:20.179060 4810 scope.go:117] "RemoveContainer" containerID="8ff6f4c3e5575c3f21c68d099e43893daf3464530dbe4d16fcaecd50a3e8a73b" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.815428 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nln5f"] Oct 08 07:06:04 crc kubenswrapper[4810]: E1008 07:06:04.816426 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="extract-utilities" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.816444 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="extract-utilities" Oct 08 07:06:04 crc kubenswrapper[4810]: E1008 07:06:04.816469 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="extract-content" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.816478 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="extract-content" Oct 08 07:06:04 crc kubenswrapper[4810]: E1008 07:06:04.816504 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="registry-server" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.816516 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="registry-server" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.816748 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9330d29-ee52-4255-bb66-2871eec3e612" containerName="registry-server" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.825334 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.827273 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nln5f"] Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.898384 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88z26\" (UniqueName: \"kubernetes.io/projected/4b819a31-f7fa-404a-882f-f418c7201e8b-kube-api-access-88z26\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.898712 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-catalog-content\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:04 crc kubenswrapper[4810]: I1008 07:06:04.898807 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-utilities\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.000707 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-catalog-content\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.001273 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-utilities\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.001389 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-catalog-content\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.001693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-utilities\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.001850 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88z26\" (UniqueName: \"kubernetes.io/projected/4b819a31-f7fa-404a-882f-f418c7201e8b-kube-api-access-88z26\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.032833 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88z26\" (UniqueName: \"kubernetes.io/projected/4b819a31-f7fa-404a-882f-f418c7201e8b-kube-api-access-88z26\") pod \"community-operators-nln5f\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.181423 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:05 crc kubenswrapper[4810]: I1008 07:06:05.692413 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nln5f"] Oct 08 07:06:06 crc kubenswrapper[4810]: I1008 07:06:06.523881 4810 generic.go:334] "Generic (PLEG): container finished" podID="4b819a31-f7fa-404a-882f-f418c7201e8b" containerID="5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480" exitCode=0 Oct 08 07:06:06 crc kubenswrapper[4810]: I1008 07:06:06.523932 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nln5f" event={"ID":"4b819a31-f7fa-404a-882f-f418c7201e8b","Type":"ContainerDied","Data":"5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480"} Oct 08 07:06:06 crc kubenswrapper[4810]: I1008 07:06:06.524228 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nln5f" event={"ID":"4b819a31-f7fa-404a-882f-f418c7201e8b","Type":"ContainerStarted","Data":"b03be3d94be1b5fdb019d541039bfa3e8a7a8748f64c8ce60ed5a7327b7dc72a"} Oct 08 07:06:07 crc kubenswrapper[4810]: I1008 07:06:07.533585 4810 generic.go:334] "Generic (PLEG): container finished" podID="4b819a31-f7fa-404a-882f-f418c7201e8b" containerID="d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f" exitCode=0 Oct 08 07:06:07 crc kubenswrapper[4810]: I1008 07:06:07.533668 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nln5f" event={"ID":"4b819a31-f7fa-404a-882f-f418c7201e8b","Type":"ContainerDied","Data":"d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f"} Oct 08 07:06:08 crc kubenswrapper[4810]: I1008 07:06:08.549446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nln5f" event={"ID":"4b819a31-f7fa-404a-882f-f418c7201e8b","Type":"ContainerStarted","Data":"f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b"} Oct 08 07:06:08 crc kubenswrapper[4810]: I1008 07:06:08.575750 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nln5f" podStartSLOduration=3.161543623 podStartE2EDuration="4.575716925s" podCreationTimestamp="2025-10-08 07:06:04 +0000 UTC" firstStartedPulling="2025-10-08 07:06:06.526009256 +0000 UTC m=+2069.160448996" lastFinishedPulling="2025-10-08 07:06:07.940182558 +0000 UTC m=+2070.574622298" observedRunningTime="2025-10-08 07:06:08.569361869 +0000 UTC m=+2071.203801619" watchObservedRunningTime="2025-10-08 07:06:08.575716925 +0000 UTC m=+2071.210156705" Oct 08 07:06:15 crc kubenswrapper[4810]: I1008 07:06:15.182234 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:15 crc kubenswrapper[4810]: I1008 07:06:15.182672 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:15 crc kubenswrapper[4810]: I1008 07:06:15.222756 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:15 crc kubenswrapper[4810]: I1008 07:06:15.643659 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:15 crc kubenswrapper[4810]: I1008 07:06:15.697731 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nln5f"] Oct 08 07:06:17 crc kubenswrapper[4810]: I1008 07:06:17.623124 4810 generic.go:334] "Generic (PLEG): container finished" podID="c7168350-094e-408e-9e5f-472928c39636" containerID="96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141" exitCode=0 Oct 08 07:06:17 crc kubenswrapper[4810]: I1008 07:06:17.623304 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" event={"ID":"c7168350-094e-408e-9e5f-472928c39636","Type":"ContainerDied","Data":"96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141"} Oct 08 07:06:17 crc kubenswrapper[4810]: I1008 07:06:17.623530 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nln5f" podUID="4b819a31-f7fa-404a-882f-f418c7201e8b" containerName="registry-server" containerID="cri-o://f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b" gracePeriod=2 Oct 08 07:06:17 crc kubenswrapper[4810]: I1008 07:06:17.624506 4810 scope.go:117] "RemoveContainer" containerID="96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141" Oct 08 07:06:17 crc kubenswrapper[4810]: I1008 07:06:17.907799 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kzxbp_must-gather-2zfqm_c7168350-094e-408e-9e5f-472928c39636/gather/0.log" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.160439 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.215835 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88z26\" (UniqueName: \"kubernetes.io/projected/4b819a31-f7fa-404a-882f-f418c7201e8b-kube-api-access-88z26\") pod \"4b819a31-f7fa-404a-882f-f418c7201e8b\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.215999 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-utilities\") pod \"4b819a31-f7fa-404a-882f-f418c7201e8b\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.216265 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-catalog-content\") pod \"4b819a31-f7fa-404a-882f-f418c7201e8b\" (UID: \"4b819a31-f7fa-404a-882f-f418c7201e8b\") " Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.217425 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-utilities" (OuterVolumeSpecName: "utilities") pod "4b819a31-f7fa-404a-882f-f418c7201e8b" (UID: "4b819a31-f7fa-404a-882f-f418c7201e8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.227362 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b819a31-f7fa-404a-882f-f418c7201e8b-kube-api-access-88z26" (OuterVolumeSpecName: "kube-api-access-88z26") pod "4b819a31-f7fa-404a-882f-f418c7201e8b" (UID: "4b819a31-f7fa-404a-882f-f418c7201e8b"). InnerVolumeSpecName "kube-api-access-88z26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.267757 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b819a31-f7fa-404a-882f-f418c7201e8b" (UID: "4b819a31-f7fa-404a-882f-f418c7201e8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.318933 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.318999 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88z26\" (UniqueName: \"kubernetes.io/projected/4b819a31-f7fa-404a-882f-f418c7201e8b-kube-api-access-88z26\") on node \"crc\" DevicePath \"\"" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.319020 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b819a31-f7fa-404a-882f-f418c7201e8b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.635032 4810 generic.go:334] "Generic (PLEG): container finished" podID="4b819a31-f7fa-404a-882f-f418c7201e8b" containerID="f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b" exitCode=0 Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.635105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nln5f" event={"ID":"4b819a31-f7fa-404a-882f-f418c7201e8b","Type":"ContainerDied","Data":"f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b"} Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.635179 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nln5f" event={"ID":"4b819a31-f7fa-404a-882f-f418c7201e8b","Type":"ContainerDied","Data":"b03be3d94be1b5fdb019d541039bfa3e8a7a8748f64c8ce60ed5a7327b7dc72a"} Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.635180 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nln5f" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.635204 4810 scope.go:117] "RemoveContainer" containerID="f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.663223 4810 scope.go:117] "RemoveContainer" containerID="d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.690524 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nln5f"] Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.694902 4810 scope.go:117] "RemoveContainer" containerID="5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.698825 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nln5f"] Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.712824 4810 scope.go:117] "RemoveContainer" containerID="f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b" Oct 08 07:06:18 crc kubenswrapper[4810]: E1008 07:06:18.713738 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b\": container with ID starting with f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b not found: ID does not exist" containerID="f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.713773 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b"} err="failed to get container status \"f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b\": rpc error: code = NotFound desc = could not find container \"f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b\": container with ID starting with f48d629affbe5cd8321b2b488aaf205a7ec4cd9cd545daee6a885f9b1a35ab4b not found: ID does not exist" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.713799 4810 scope.go:117] "RemoveContainer" containerID="d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f" Oct 08 07:06:18 crc kubenswrapper[4810]: E1008 07:06:18.714253 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f\": container with ID starting with d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f not found: ID does not exist" containerID="d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.714302 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f"} err="failed to get container status \"d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f\": rpc error: code = NotFound desc = could not find container \"d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f\": container with ID starting with d1ff97eafc4ff3ae8f31a1df1cfaf7a798618194c5dae5520731ea870034622f not found: ID does not exist" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.714332 4810 scope.go:117] "RemoveContainer" containerID="5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480" Oct 08 07:06:18 crc kubenswrapper[4810]: E1008 07:06:18.715361 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480\": container with ID starting with 5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480 not found: ID does not exist" containerID="5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480" Oct 08 07:06:18 crc kubenswrapper[4810]: I1008 07:06:18.715392 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480"} err="failed to get container status \"5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480\": rpc error: code = NotFound desc = could not find container \"5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480\": container with ID starting with 5dfeb14781e27e5a4cd6f2f1d9de9f91b540dc262b571b547cb4fbc7391fb480 not found: ID does not exist" Oct 08 07:06:20 crc kubenswrapper[4810]: I1008 07:06:20.081219 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b819a31-f7fa-404a-882f-f418c7201e8b" path="/var/lib/kubelet/pods/4b819a31-f7fa-404a-882f-f418c7201e8b/volumes" Oct 08 07:06:25 crc kubenswrapper[4810]: I1008 07:06:25.966333 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kzxbp/must-gather-2zfqm"] Oct 08 07:06:25 crc kubenswrapper[4810]: I1008 07:06:25.967149 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" podUID="c7168350-094e-408e-9e5f-472928c39636" containerName="copy" containerID="cri-o://9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838" gracePeriod=2 Oct 08 07:06:25 crc kubenswrapper[4810]: I1008 07:06:25.972564 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kzxbp/must-gather-2zfqm"] Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.409341 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kzxbp_must-gather-2zfqm_c7168350-094e-408e-9e5f-472928c39636/copy/0.log" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.410327 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.449623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffq2j\" (UniqueName: \"kubernetes.io/projected/c7168350-094e-408e-9e5f-472928c39636-kube-api-access-ffq2j\") pod \"c7168350-094e-408e-9e5f-472928c39636\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.449742 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7168350-094e-408e-9e5f-472928c39636-must-gather-output\") pod \"c7168350-094e-408e-9e5f-472928c39636\" (UID: \"c7168350-094e-408e-9e5f-472928c39636\") " Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.455803 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7168350-094e-408e-9e5f-472928c39636-kube-api-access-ffq2j" (OuterVolumeSpecName: "kube-api-access-ffq2j") pod "c7168350-094e-408e-9e5f-472928c39636" (UID: "c7168350-094e-408e-9e5f-472928c39636"). InnerVolumeSpecName "kube-api-access-ffq2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.551142 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffq2j\" (UniqueName: \"kubernetes.io/projected/c7168350-094e-408e-9e5f-472928c39636-kube-api-access-ffq2j\") on node \"crc\" DevicePath \"\"" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.553833 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7168350-094e-408e-9e5f-472928c39636-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c7168350-094e-408e-9e5f-472928c39636" (UID: "c7168350-094e-408e-9e5f-472928c39636"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.651938 4810 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7168350-094e-408e-9e5f-472928c39636-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.705225 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kzxbp_must-gather-2zfqm_c7168350-094e-408e-9e5f-472928c39636/copy/0.log" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.706083 4810 generic.go:334] "Generic (PLEG): container finished" podID="c7168350-094e-408e-9e5f-472928c39636" containerID="9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838" exitCode=143 Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.706144 4810 scope.go:117] "RemoveContainer" containerID="9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.706173 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kzxbp/must-gather-2zfqm" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.739396 4810 scope.go:117] "RemoveContainer" containerID="96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.801244 4810 scope.go:117] "RemoveContainer" containerID="9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838" Oct 08 07:06:26 crc kubenswrapper[4810]: E1008 07:06:26.801720 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838\": container with ID starting with 9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838 not found: ID does not exist" containerID="9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.801776 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838"} err="failed to get container status \"9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838\": rpc error: code = NotFound desc = could not find container \"9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838\": container with ID starting with 9094cb691c64ce36b30798b8b2620866e4db090babcde3c3d8abeec58c2fe838 not found: ID does not exist" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.801805 4810 scope.go:117] "RemoveContainer" containerID="96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141" Oct 08 07:06:26 crc kubenswrapper[4810]: E1008 07:06:26.802398 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141\": container with ID starting with 96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141 not found: ID does not exist" containerID="96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141" Oct 08 07:06:26 crc kubenswrapper[4810]: I1008 07:06:26.802423 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141"} err="failed to get container status \"96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141\": rpc error: code = NotFound desc = could not find container \"96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141\": container with ID starting with 96a5b29f87fff75c0a0b3912c42227e6af3a9e82e96c2f52525149ccf4725141 not found: ID does not exist" Oct 08 07:06:28 crc kubenswrapper[4810]: I1008 07:06:28.086310 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7168350-094e-408e-9e5f-472928c39636" path="/var/lib/kubelet/pods/c7168350-094e-408e-9e5f-472928c39636/volumes" Oct 08 07:07:19 crc kubenswrapper[4810]: I1008 07:07:19.431935 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 07:07:19 crc kubenswrapper[4810]: I1008 07:07:19.432654 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 07:07:49 crc kubenswrapper[4810]: I1008 07:07:49.432182 4810 patch_prober.go:28] interesting pod/machine-config-daemon-7vlk5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 07:07:49 crc kubenswrapper[4810]: I1008 07:07:49.432736 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7vlk5" podUID="2b5ec8eb-9a64-46ae-b881-535323e3b686" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071406742024453 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071406742017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071402133016501 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071402134015452 5ustar corecore